mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-26 19:01:44 +02:00
a04ad248a0
This is possible now that the next-API is gone. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com> Signed-off-by: James Almer <jamrial@gmail.com>
318 lines
12 KiB
C
318 lines
12 KiB
C
/*
|
|
* Copyright (c) 2002 Michael Niedermayer <michaelni@gmx.at>
|
|
* Copyright (c) 2011 Stefano Sabatini
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with FFmpeg; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* Apply a boxblur filter to the input video.
|
|
* Ported from MPlayer libmpcodecs/vf_boxblur.c.
|
|
*/
|
|
|
|
#include "libavutil/avstring.h"
|
|
#include "libavutil/common.h"
|
|
#include "libavutil/opt.h"
|
|
#include "avfilter.h"
|
|
#include "formats.h"
|
|
#include "internal.h"
|
|
#include "video.h"
|
|
#include "boxblur.h"
|
|
|
|
|
|
typedef struct BoxBlurContext {
|
|
const AVClass *class;
|
|
FilterParam luma_param;
|
|
FilterParam chroma_param;
|
|
FilterParam alpha_param;
|
|
|
|
int hsub, vsub;
|
|
int radius[4];
|
|
int power[4];
|
|
uint8_t *temp[2]; ///< temporary buffer used in blur_power()
|
|
} BoxBlurContext;
|
|
|
|
static av_cold void uninit(AVFilterContext *ctx)
|
|
{
|
|
BoxBlurContext *s = ctx->priv;
|
|
|
|
av_freep(&s->temp[0]);
|
|
av_freep(&s->temp[1]);
|
|
}
|
|
|
|
static int query_formats(AVFilterContext *ctx)
|
|
{
|
|
AVFilterFormats *formats = NULL;
|
|
int fmt, ret;
|
|
|
|
for (fmt = 0; av_pix_fmt_desc_get(fmt); fmt++) {
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(fmt);
|
|
if (!(desc->flags & (AV_PIX_FMT_FLAG_HWACCEL | AV_PIX_FMT_FLAG_BITSTREAM | AV_PIX_FMT_FLAG_PAL)) &&
|
|
(desc->flags & AV_PIX_FMT_FLAG_PLANAR || desc->nb_components == 1) &&
|
|
(!(desc->flags & AV_PIX_FMT_FLAG_BE) == !HAVE_BIGENDIAN || desc->comp[0].depth == 8) &&
|
|
(ret = ff_add_format(&formats, fmt)) < 0)
|
|
return ret;
|
|
}
|
|
|
|
return ff_set_common_formats(ctx, formats);
|
|
}
|
|
|
|
static int config_input(AVFilterLink *inlink)
|
|
{
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
|
|
AVFilterContext *ctx = inlink->dst;
|
|
BoxBlurContext *s = ctx->priv;
|
|
int w = inlink->w, h = inlink->h;
|
|
int ret;
|
|
|
|
if (!(s->temp[0] = av_malloc(2*FFMAX(w, h))) ||
|
|
!(s->temp[1] = av_malloc(2*FFMAX(w, h))))
|
|
return AVERROR(ENOMEM);
|
|
|
|
s->hsub = desc->log2_chroma_w;
|
|
s->vsub = desc->log2_chroma_h;
|
|
|
|
ret = ff_boxblur_eval_filter_params(inlink,
|
|
&s->luma_param,
|
|
&s->chroma_param,
|
|
&s->alpha_param);
|
|
|
|
if (ret != 0) {
|
|
av_log(ctx, AV_LOG_ERROR, "Failed to evaluate "
|
|
"filter params: %d.\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
s->radius[Y] = s->luma_param.radius;
|
|
s->radius[U] = s->radius[V] = s->chroma_param.radius;
|
|
s->radius[A] = s->alpha_param.radius;
|
|
|
|
s->power[Y] = s->luma_param.power;
|
|
s->power[U] = s->power[V] = s->chroma_param.power;
|
|
s->power[A] = s->alpha_param.power;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Naive boxblur would sum source pixels from x-radius .. x+radius
|
|
* for destination pixel x. That would be O(radius*width).
|
|
* If you now look at what source pixels represent 2 consecutive
|
|
* output pixels, then you see they are almost identical and only
|
|
* differ by 2 pixels, like:
|
|
* src0 111111111
|
|
* dst0 1
|
|
* src1 111111111
|
|
* dst1 1
|
|
* src0-src1 1 -1
|
|
* so when you know one output pixel you can find the next by just adding
|
|
* and subtracting 1 input pixel.
|
|
* The following code adopts this faster variant.
|
|
*/
|
|
#define BLUR(type, depth) \
|
|
static inline void blur ## depth(type *dst, int dst_step, const type *src, \
|
|
int src_step, int len, int radius) \
|
|
{ \
|
|
const int length = radius*2 + 1; \
|
|
const int inv = ((1<<16) + length/2)/length; \
|
|
int x, sum = src[radius*src_step]; \
|
|
\
|
|
for (x = 0; x < radius; x++) \
|
|
sum += src[x*src_step]<<1; \
|
|
\
|
|
sum = sum*inv + (1<<15); \
|
|
\
|
|
for (x = 0; x <= radius; x++) { \
|
|
sum += (src[(radius+x)*src_step] - src[(radius-x)*src_step])*inv; \
|
|
dst[x*dst_step] = sum>>16; \
|
|
} \
|
|
\
|
|
for (; x < len-radius; x++) { \
|
|
sum += (src[(radius+x)*src_step] - src[(x-radius-1)*src_step])*inv; \
|
|
dst[x*dst_step] = sum >>16; \
|
|
} \
|
|
\
|
|
for (; x < len; x++) { \
|
|
sum += (src[(2*len-radius-x-1)*src_step] - src[(x-radius-1)*src_step])*inv; \
|
|
dst[x*dst_step] = sum>>16; \
|
|
} \
|
|
}
|
|
|
|
BLUR(uint8_t, 8)
|
|
BLUR(uint16_t, 16)
|
|
|
|
#undef BLUR
|
|
|
|
static inline void blur(uint8_t *dst, int dst_step, const uint8_t *src, int src_step,
|
|
int len, int radius, int pixsize)
|
|
{
|
|
if (pixsize == 1) blur8 (dst, dst_step , src, src_step , len, radius);
|
|
else blur16((uint16_t*)dst, dst_step>>1, (const uint16_t*)src, src_step>>1, len, radius);
|
|
}
|
|
|
|
static inline void blur_power(uint8_t *dst, int dst_step, const uint8_t *src, int src_step,
|
|
int len, int radius, int power, uint8_t *temp[2], int pixsize)
|
|
{
|
|
uint8_t *a = temp[0], *b = temp[1];
|
|
|
|
if (radius && power) {
|
|
blur(a, pixsize, src, src_step, len, radius, pixsize);
|
|
for (; power > 2; power--) {
|
|
uint8_t *c;
|
|
blur(b, pixsize, a, pixsize, len, radius, pixsize);
|
|
c = a; a = b; b = c;
|
|
}
|
|
if (power > 1) {
|
|
blur(dst, dst_step, a, pixsize, len, radius, pixsize);
|
|
} else {
|
|
int i;
|
|
if (pixsize == 1) {
|
|
for (i = 0; i < len; i++)
|
|
dst[i*dst_step] = a[i];
|
|
} else
|
|
for (i = 0; i < len; i++)
|
|
*(uint16_t*)(dst + i*dst_step) = ((uint16_t*)a)[i];
|
|
}
|
|
} else {
|
|
int i;
|
|
if (pixsize == 1) {
|
|
for (i = 0; i < len; i++)
|
|
dst[i*dst_step] = src[i*src_step];
|
|
} else
|
|
for (i = 0; i < len; i++)
|
|
*(uint16_t*)(dst + i*dst_step) = *(uint16_t*)(src + i*src_step);
|
|
}
|
|
}
|
|
|
|
static void hblur(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize,
|
|
int w, int h, int radius, int power, uint8_t *temp[2], int pixsize)
|
|
{
|
|
int y;
|
|
|
|
if (radius == 0 && dst == src)
|
|
return;
|
|
|
|
for (y = 0; y < h; y++)
|
|
blur_power(dst + y*dst_linesize, pixsize, src + y*src_linesize, pixsize,
|
|
w, radius, power, temp, pixsize);
|
|
}
|
|
|
|
static void vblur(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize,
|
|
int w, int h, int radius, int power, uint8_t *temp[2], int pixsize)
|
|
{
|
|
int x;
|
|
|
|
if (radius == 0 && dst == src)
|
|
return;
|
|
|
|
for (x = 0; x < w; x++)
|
|
blur_power(dst + x*pixsize, dst_linesize, src + x*pixsize, src_linesize,
|
|
h, radius, power, temp, pixsize);
|
|
}
|
|
|
|
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
|
|
{
|
|
AVFilterContext *ctx = inlink->dst;
|
|
BoxBlurContext *s = ctx->priv;
|
|
AVFilterLink *outlink = inlink->dst->outputs[0];
|
|
AVFrame *out;
|
|
int plane;
|
|
int cw = AV_CEIL_RSHIFT(inlink->w, s->hsub), ch = AV_CEIL_RSHIFT(in->height, s->vsub);
|
|
int w[4] = { inlink->w, cw, cw, inlink->w };
|
|
int h[4] = { in->height, ch, ch, in->height };
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
|
|
const int depth = desc->comp[0].depth;
|
|
const int pixsize = (depth+7)/8;
|
|
|
|
out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
|
|
if (!out) {
|
|
av_frame_free(&in);
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
av_frame_copy_props(out, in);
|
|
|
|
for (plane = 0; plane < 4 && in->data[plane] && in->linesize[plane]; plane++)
|
|
hblur(out->data[plane], out->linesize[plane],
|
|
in ->data[plane], in ->linesize[plane],
|
|
w[plane], h[plane], s->radius[plane], s->power[plane],
|
|
s->temp, pixsize);
|
|
|
|
for (plane = 0; plane < 4 && in->data[plane] && in->linesize[plane]; plane++)
|
|
vblur(out->data[plane], out->linesize[plane],
|
|
out->data[plane], out->linesize[plane],
|
|
w[plane], h[plane], s->radius[plane], s->power[plane],
|
|
s->temp, pixsize);
|
|
|
|
av_frame_free(&in);
|
|
|
|
return ff_filter_frame(outlink, out);
|
|
}
|
|
|
|
#define OFFSET(x) offsetof(BoxBlurContext, x)
|
|
#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
|
|
|
|
static const AVOption boxblur_options[] = {
|
|
{ "luma_radius", "Radius of the luma blurring box", OFFSET(luma_param.radius_expr), AV_OPT_TYPE_STRING, {.str="2"}, .flags = FLAGS },
|
|
{ "lr", "Radius of the luma blurring box", OFFSET(luma_param.radius_expr), AV_OPT_TYPE_STRING, {.str="2"}, .flags = FLAGS },
|
|
{ "luma_power", "How many times should the boxblur be applied to luma", OFFSET(luma_param.power), AV_OPT_TYPE_INT, {.i64=2}, 0, INT_MAX, .flags = FLAGS },
|
|
{ "lp", "How many times should the boxblur be applied to luma", OFFSET(luma_param.power), AV_OPT_TYPE_INT, {.i64=2}, 0, INT_MAX, .flags = FLAGS },
|
|
|
|
{ "chroma_radius", "Radius of the chroma blurring box", OFFSET(chroma_param.radius_expr), AV_OPT_TYPE_STRING, {.str=NULL}, .flags = FLAGS },
|
|
{ "cr", "Radius of the chroma blurring box", OFFSET(chroma_param.radius_expr), AV_OPT_TYPE_STRING, {.str=NULL}, .flags = FLAGS },
|
|
{ "chroma_power", "How many times should the boxblur be applied to chroma", OFFSET(chroma_param.power), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, .flags = FLAGS },
|
|
{ "cp", "How many times should the boxblur be applied to chroma", OFFSET(chroma_param.power), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, .flags = FLAGS },
|
|
|
|
{ "alpha_radius", "Radius of the alpha blurring box", OFFSET(alpha_param.radius_expr), AV_OPT_TYPE_STRING, {.str=NULL}, .flags = FLAGS },
|
|
{ "ar", "Radius of the alpha blurring box", OFFSET(alpha_param.radius_expr), AV_OPT_TYPE_STRING, {.str=NULL}, .flags = FLAGS },
|
|
{ "alpha_power", "How many times should the boxblur be applied to alpha", OFFSET(alpha_param.power), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, .flags = FLAGS },
|
|
{ "ap", "How many times should the boxblur be applied to alpha", OFFSET(alpha_param.power), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, .flags = FLAGS },
|
|
|
|
{ NULL }
|
|
};
|
|
|
|
AVFILTER_DEFINE_CLASS(boxblur);
|
|
|
|
static const AVFilterPad avfilter_vf_boxblur_inputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.config_props = config_input,
|
|
.filter_frame = filter_frame,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVFilterPad avfilter_vf_boxblur_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
const AVFilter ff_vf_boxblur = {
|
|
.name = "boxblur",
|
|
.description = NULL_IF_CONFIG_SMALL("Blur the input."),
|
|
.priv_size = sizeof(BoxBlurContext),
|
|
.priv_class = &boxblur_class,
|
|
.uninit = uninit,
|
|
.query_formats = query_formats,
|
|
.inputs = avfilter_vf_boxblur_inputs,
|
|
.outputs = avfilter_vf_boxblur_outputs,
|
|
.flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC,
|
|
};
|