mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-26 19:01:44 +02:00
314 lines
10 KiB
C
314 lines
10 KiB
C
/*
|
|
* Copyright (c) 2002 Michael Niedermayer <michaelni@gmx.at>
|
|
* Copyright (c) 2012 Jeremy Tran
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with FFmpeg; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* Apply a smartblur filter to the input video
|
|
* Ported from MPlayer libmpcodecs/vf_smartblur.c by Michael Niedermayer.
|
|
*/
|
|
|
|
#include "libavutil/pixdesc.h"
|
|
#include "libswscale/swscale.h"
|
|
|
|
#include "avfilter.h"
|
|
#include "formats.h"
|
|
#include "internal.h"
|
|
|
|
#define RADIUS_MIN 0.1
|
|
#define RADIUS_MAX 5.0
|
|
|
|
#define STRENGTH_MIN -1.0
|
|
#define STRENGTH_MAX 1.0
|
|
|
|
#define THRESHOLD_MIN -30
|
|
#define THRESHOLD_MAX 30
|
|
|
|
typedef struct {
|
|
float radius;
|
|
float strength;
|
|
int threshold;
|
|
float quality;
|
|
struct SwsContext *filter_context;
|
|
} FilterParam;
|
|
|
|
typedef struct {
|
|
FilterParam luma;
|
|
FilterParam chroma;
|
|
int hsub;
|
|
int vsub;
|
|
unsigned int sws_flags;
|
|
} SmartblurContext;
|
|
|
|
#define CHECK_PARAM(param, name, min, max, format, ret) \
|
|
if (param < min || param > max) { \
|
|
av_log(ctx, AV_LOG_ERROR, \
|
|
"Invalid " #name " value " #format ": " \
|
|
"must be included between range " #format " and " #format "\n",\
|
|
param, min, max); \
|
|
ret = AVERROR(EINVAL); \
|
|
}
|
|
|
|
static av_cold int init(AVFilterContext *ctx, const char *args)
|
|
{
|
|
SmartblurContext *sblur = ctx->priv;
|
|
int n = 0, ret = 0;
|
|
float lradius, lstrength, cradius, cstrength;
|
|
int lthreshold, cthreshold;
|
|
|
|
if (args)
|
|
n = sscanf(args, "%f:%f:%d:%f:%f:%d",
|
|
&lradius, &lstrength, <hreshold,
|
|
&cradius, &cstrength, &cthreshold);
|
|
|
|
if (n != 3 && n != 6) {
|
|
av_log(ctx, AV_LOG_ERROR,
|
|
"Incorrect number of parameters or invalid syntax: "
|
|
"must be luma_radius:luma_strength:luma_threshold"
|
|
"[:chroma_radius:chroma_strength:chroma_threshold]\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
sblur->luma.radius = lradius;
|
|
sblur->luma.strength = lstrength;
|
|
sblur->luma.threshold = lthreshold;
|
|
|
|
if (n == 3) {
|
|
sblur->chroma.radius = sblur->luma.radius;
|
|
sblur->chroma.strength = sblur->luma.strength;
|
|
sblur->chroma.threshold = sblur->luma.threshold;
|
|
} else {
|
|
sblur->chroma.radius = cradius;
|
|
sblur->chroma.strength = cstrength;
|
|
sblur->chroma.threshold = cthreshold;
|
|
}
|
|
|
|
sblur->luma.quality = sblur->chroma.quality = 3.0;
|
|
sblur->sws_flags = SWS_BICUBIC;
|
|
|
|
CHECK_PARAM(lradius, luma radius, RADIUS_MIN, RADIUS_MAX, %0.1f, ret)
|
|
CHECK_PARAM(lstrength, luma strength, STRENGTH_MIN, STRENGTH_MAX, %0.1f, ret)
|
|
CHECK_PARAM(lthreshold, luma threshold, THRESHOLD_MIN, THRESHOLD_MAX, %d, ret)
|
|
|
|
if (n != 3) {
|
|
CHECK_PARAM(sblur->chroma.radius, chroma radius, RADIUS_MIN, RADIUS_MAX, %0.1f, ret)
|
|
CHECK_PARAM(sblur->chroma.strength, chroma strength, STRENGTH_MIN, STRENGTH_MAX, %0.1f, ret)
|
|
CHECK_PARAM(sblur->chroma.threshold, chroma threshold, THRESHOLD_MIN,THRESHOLD_MAX, %d, ret)
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static av_cold void uninit(AVFilterContext *ctx)
|
|
{
|
|
SmartblurContext *sblur = ctx->priv;
|
|
|
|
sws_freeContext(sblur->luma.filter_context);
|
|
sws_freeContext(sblur->chroma.filter_context);
|
|
}
|
|
|
|
static int query_formats(AVFilterContext *ctx)
|
|
{
|
|
static const enum AVPixelFormat pix_fmts[] = {
|
|
AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P,
|
|
AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P,
|
|
AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV440P,
|
|
AV_PIX_FMT_GRAY8,
|
|
AV_PIX_FMT_NONE
|
|
};
|
|
|
|
ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int alloc_sws_context(FilterParam *f, int width, int height, unsigned int flags)
|
|
{
|
|
SwsVector *vec;
|
|
SwsFilter sws_filter;
|
|
|
|
vec = sws_getGaussianVec(f->radius, f->quality);
|
|
|
|
if (!vec)
|
|
return AVERROR(EINVAL);
|
|
|
|
sws_scaleVec(vec, f->strength);
|
|
vec->coeff[vec->length / 2] += 1.0 - f->strength;
|
|
sws_filter.lumH = sws_filter.lumV = vec;
|
|
sws_filter.chrH = sws_filter.chrV = NULL;
|
|
f->filter_context = sws_getCachedContext(NULL,
|
|
width, height, AV_PIX_FMT_GRAY8,
|
|
width, height, AV_PIX_FMT_GRAY8,
|
|
flags, &sws_filter, NULL, NULL);
|
|
|
|
sws_freeVec(vec);
|
|
|
|
if (!f->filter_context)
|
|
return AVERROR(EINVAL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int config_props(AVFilterLink *inlink)
|
|
{
|
|
SmartblurContext *sblur = inlink->dst->priv;
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
|
|
|
|
sblur->hsub = desc->log2_chroma_w;
|
|
sblur->vsub = desc->log2_chroma_h;
|
|
|
|
alloc_sws_context(&sblur->luma, inlink->w, inlink->h, sblur->sws_flags);
|
|
alloc_sws_context(&sblur->chroma,
|
|
inlink->w >> sblur->hsub, inlink->h >> sblur->vsub,
|
|
sblur->sws_flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void blur(uint8_t *dst, const int dst_linesize,
|
|
const uint8_t *src, const int src_linesize,
|
|
const int w, const int h, const int threshold,
|
|
struct SwsContext *filter_context)
|
|
{
|
|
int x, y;
|
|
int orig, filtered;
|
|
int diff;
|
|
/* Declare arrays of 4 to get aligned data */
|
|
const uint8_t* const src_array[4] = {src};
|
|
uint8_t *dst_array[4] = {dst};
|
|
int src_linesize_array[4] = {src_linesize};
|
|
int dst_linesize_array[4] = {dst_linesize};
|
|
|
|
sws_scale(filter_context, src_array, src_linesize_array,
|
|
0, h, dst_array, dst_linesize_array);
|
|
|
|
if (threshold > 0) {
|
|
for (y = 0; y < h; ++y) {
|
|
for (x = 0; x < w; ++x) {
|
|
orig = src[x + y * src_linesize];
|
|
filtered = dst[x + y * dst_linesize];
|
|
diff = orig - filtered;
|
|
|
|
if (diff > 0) {
|
|
if (diff > 2 * threshold)
|
|
dst[x + y * dst_linesize] = orig;
|
|
else if (diff > threshold)
|
|
/* add 'diff' and substract 'threshold' from 'filtered' */
|
|
dst[x + y * dst_linesize] = orig - threshold;
|
|
} else {
|
|
if (-diff > 2 * threshold)
|
|
dst[x + y * dst_linesize] = orig;
|
|
else if (-diff > threshold)
|
|
/* add 'diff' and 'threshold' to 'filtered' */
|
|
dst[x + y * dst_linesize] = orig + threshold;
|
|
}
|
|
}
|
|
}
|
|
} else if (threshold < 0) {
|
|
for (y = 0; y < h; ++y) {
|
|
for (x = 0; x < w; ++x) {
|
|
orig = src[x + y * src_linesize];
|
|
filtered = dst[x + y * dst_linesize];
|
|
diff = orig - filtered;
|
|
|
|
if (diff > 0) {
|
|
if (diff <= -threshold)
|
|
dst[x + y * dst_linesize] = orig;
|
|
else if (diff <= -2 * threshold)
|
|
/* substract 'diff' and 'threshold' from 'orig' */
|
|
dst[x + y * dst_linesize] = filtered - threshold;
|
|
} else {
|
|
if (diff >= threshold)
|
|
dst[x + y * dst_linesize] = orig;
|
|
else if (diff >= 2 * threshold)
|
|
/* add 'threshold' and substract 'diff' from 'orig' */
|
|
dst[x + y * dst_linesize] = filtered + threshold;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
static int filter_frame(AVFilterLink *inlink, AVFrame *inpic)
|
|
{
|
|
SmartblurContext *sblur = inlink->dst->priv;
|
|
AVFilterLink *outlink = inlink->dst->outputs[0];
|
|
AVFrame *outpic;
|
|
int cw = inlink->w >> sblur->hsub;
|
|
int ch = inlink->h >> sblur->vsub;
|
|
|
|
outpic = ff_get_video_buffer(outlink, outlink->w, outlink->h);
|
|
if (!outpic) {
|
|
av_frame_free(&inpic);
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
av_frame_copy_props(outpic, inpic);
|
|
|
|
blur(outpic->data[0], outpic->linesize[0],
|
|
inpic->data[0], inpic->linesize[0],
|
|
inlink->w, inlink->h, sblur->luma.threshold,
|
|
sblur->luma.filter_context);
|
|
|
|
if (inpic->data[2]) {
|
|
blur(outpic->data[1], outpic->linesize[1],
|
|
inpic->data[1], inpic->linesize[1],
|
|
cw, ch, sblur->chroma.threshold,
|
|
sblur->chroma.filter_context);
|
|
blur(outpic->data[2], outpic->linesize[2],
|
|
inpic->data[2], inpic->linesize[2],
|
|
cw, ch, sblur->chroma.threshold,
|
|
sblur->chroma.filter_context);
|
|
}
|
|
|
|
av_frame_free(&inpic);
|
|
return ff_filter_frame(outlink, outpic);
|
|
}
|
|
|
|
static const AVFilterPad smartblur_inputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.filter_frame = filter_frame,
|
|
.config_props = config_props,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVFilterPad smartblur_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
AVFilter avfilter_vf_smartblur = {
|
|
.name = "smartblur",
|
|
.description = NULL_IF_CONFIG_SMALL("Blur the input video without impacting the outlines."),
|
|
|
|
.priv_size = sizeof(SmartblurContext),
|
|
|
|
.init = init,
|
|
.uninit = uninit,
|
|
.query_formats = query_formats,
|
|
.inputs = smartblur_inputs,
|
|
.outputs = smartblur_outputs,
|
|
};
|