1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00
FFmpeg/libavfilter/af_silenceremove.c
Andreas Rheinhardt 790f793844 avutil/common: Don't auto-include mem.h
There are lots of files that don't need it: The number of object
files that actually need it went down from 2011 to 884 here.

Keep it for external users in order to not cause breakages.

Also improve the other headers a bit while just at it.

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
2024-03-31 00:08:43 +01:00

497 lines
18 KiB
C

/*
* Copyright (c) 2001 Heikki Leinonen
* Copyright (c) 2001 Chris Bagwell
* Copyright (c) 2003 Donnie Smith
* Copyright (c) 2014 Paul B Mahol
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <float.h> /* DBL_MAX */
#include "libavutil/avassert.h"
#include "libavutil/mem.h"
#include "libavutil/opt.h"
#include "audio.h"
#include "filters.h"
#include "avfilter.h"
#include "internal.h"
enum SilenceDetect {
D_AVG,
D_RMS,
D_PEAK,
D_MEDIAN,
D_PTP,
D_DEV,
D_NB
};
enum TimestampMode {
TS_WRITE,
TS_COPY,
TS_NB
};
enum ThresholdMode {
T_ANY,
T_ALL,
};
typedef struct SilenceRemoveContext {
const AVClass *class;
int start_mode;
int start_periods;
int64_t start_duration;
int64_t start_duration_opt;
double start_threshold;
int64_t start_silence;
int64_t start_silence_opt;
int stop_mode;
int stop_periods;
int64_t stop_duration;
int64_t stop_duration_opt;
double stop_threshold;
int64_t stop_silence;
int64_t stop_silence_opt;
int64_t window_duration_opt;
int timestamp_mode;
int start_found_periods;
int stop_found_periods;
int start_sample_count;
int start_silence_count;
int stop_sample_count;
int stop_silence_count;
AVFrame *start_window;
AVFrame *stop_window;
int *start_front;
int *start_back;
int *stop_front;
int *stop_back;
int64_t window_duration;
int cache_size;
int start_window_pos;
int start_window_size;
int stop_window_pos;
int stop_window_size;
double *start_cache;
double *stop_cache;
AVFrame *start_queuef;
int start_queue_pos;
int start_queue_size;
AVFrame *stop_queuef;
int stop_queue_pos;
int stop_queue_size;
int restart;
int found_nonsilence;
int64_t next_pts;
int detection;
float (*compute_flt)(float *c, float s, float ws, int size, int *front, int *back);
double (*compute_dbl)(double *c, double s, double ws, int size, int *front, int *back);
} SilenceRemoveContext;
#define OFFSET(x) offsetof(SilenceRemoveContext, x)
#define AF AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_AUDIO_PARAM
#define AFR AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_RUNTIME_PARAM
static const AVOption silenceremove_options[] = {
{ "start_periods", "set periods of silence parts to skip from start", OFFSET(start_periods), AV_OPT_TYPE_INT, {.i64=0}, 0, 9000, AF },
{ "start_duration", "set start duration of non-silence part", OFFSET(start_duration_opt), AV_OPT_TYPE_DURATION, {.i64=0}, 0, INT32_MAX, AF },
{ "start_threshold", "set threshold for start silence detection", OFFSET(start_threshold), AV_OPT_TYPE_DOUBLE, {.dbl=0}, 0, DBL_MAX, AFR },
{ "start_silence", "set start duration of silence part to keep", OFFSET(start_silence_opt), AV_OPT_TYPE_DURATION, {.i64=0}, 0, INT32_MAX, AF },
{ "start_mode", "set which channel will trigger trimming from start", OFFSET(start_mode), AV_OPT_TYPE_INT, {.i64=T_ANY}, T_ANY, T_ALL, AFR, .unit = "mode" },
{ "any", 0, 0, AV_OPT_TYPE_CONST, {.i64=T_ANY}, 0, 0, AFR, .unit = "mode" },
{ "all", 0, 0, AV_OPT_TYPE_CONST, {.i64=T_ALL}, 0, 0, AFR, .unit = "mode" },
{ "stop_periods", "set periods of silence parts to skip from end", OFFSET(stop_periods), AV_OPT_TYPE_INT, {.i64=0}, -9000, 9000, AF },
{ "stop_duration", "set stop duration of silence part", OFFSET(stop_duration_opt), AV_OPT_TYPE_DURATION, {.i64=0}, 0, INT32_MAX, AF },
{ "stop_threshold", "set threshold for stop silence detection", OFFSET(stop_threshold), AV_OPT_TYPE_DOUBLE, {.dbl=0}, 0, DBL_MAX, AFR },
{ "stop_silence", "set stop duration of silence part to keep", OFFSET(stop_silence_opt), AV_OPT_TYPE_DURATION, {.i64=0}, 0, INT32_MAX, AF },
{ "stop_mode", "set which channel will trigger trimming from end", OFFSET(stop_mode), AV_OPT_TYPE_INT, {.i64=T_ALL}, T_ANY, T_ALL, AFR, .unit = "mode" },
{ "detection", "set how silence is detected", OFFSET(detection), AV_OPT_TYPE_INT, {.i64=D_RMS}, 0, D_NB-1, AF, .unit = "detection" },
{ "avg", "use mean absolute values of samples", 0, AV_OPT_TYPE_CONST, {.i64=D_AVG}, 0, 0, AF, .unit = "detection" },
{ "rms", "use root mean squared values of samples", 0, AV_OPT_TYPE_CONST, {.i64=D_RMS}, 0, 0, AF, .unit = "detection" },
{ "peak", "use max absolute values of samples", 0, AV_OPT_TYPE_CONST, {.i64=D_PEAK},0, 0, AF, .unit = "detection" },
{ "median", "use median of absolute values of samples", 0, AV_OPT_TYPE_CONST, {.i64=D_MEDIAN},0, 0, AF, .unit = "detection" },
{ "ptp", "use absolute of max peak to min peak difference", 0, AV_OPT_TYPE_CONST, {.i64=D_PTP}, 0, 0, AF, .unit = "detection" },
{ "dev", "use standard deviation from values of samples", 0, AV_OPT_TYPE_CONST, {.i64=D_DEV}, 0, 0, AF, .unit = "detection" },
{ "window", "set duration of window for silence detection", OFFSET(window_duration_opt), AV_OPT_TYPE_DURATION, {.i64=20000}, 0, 100000000, AF },
{ "timestamp", "set how every output frame timestamp is processed", OFFSET(timestamp_mode), AV_OPT_TYPE_INT, {.i64=TS_WRITE}, 0, TS_NB-1, AF, .unit = "timestamp" },
{ "write", "full timestamps rewrite, keep only the start time", 0, AV_OPT_TYPE_CONST, {.i64=TS_WRITE}, 0, 0, AF, .unit = "timestamp" },
{ "copy", "non-dropped frames are left with same timestamp", 0, AV_OPT_TYPE_CONST, {.i64=TS_COPY}, 0, 0, AF, .unit = "timestamp" },
{ NULL }
};
AVFILTER_DEFINE_CLASS(silenceremove);
#define DEPTH 32
#include "silenceremove_template.c"
#undef DEPTH
#define DEPTH 64
#include "silenceremove_template.c"
static av_cold int init(AVFilterContext *ctx)
{
SilenceRemoveContext *s = ctx->priv;
if (s->stop_periods < 0) {
s->stop_periods = -s->stop_periods;
s->restart = 1;
}
return 0;
}
static void clear_windows(SilenceRemoveContext *s)
{
av_samples_set_silence(s->start_window->extended_data, 0,
s->start_window->nb_samples,
s->start_window->ch_layout.nb_channels,
s->start_window->format);
av_samples_set_silence(s->stop_window->extended_data, 0,
s->stop_window->nb_samples,
s->stop_window->ch_layout.nb_channels,
s->stop_window->format);
s->start_window_pos = 0;
s->start_window_size = 0;
s->stop_window_pos = 0;
s->stop_window_size = 0;
s->start_queue_pos = 0;
s->start_queue_size = 0;
s->stop_queue_pos = 0;
s->stop_queue_size = 0;
}
static int config_input(AVFilterLink *inlink)
{
AVFilterContext *ctx = inlink->dst;
SilenceRemoveContext *s = ctx->priv;
s->next_pts = AV_NOPTS_VALUE;
s->window_duration = av_rescale(s->window_duration_opt, inlink->sample_rate,
AV_TIME_BASE);
s->window_duration = FFMAX(1, s->window_duration);
s->start_duration = av_rescale(s->start_duration_opt, inlink->sample_rate,
AV_TIME_BASE);
s->start_silence = av_rescale(s->start_silence_opt, inlink->sample_rate,
AV_TIME_BASE);
s->stop_duration = av_rescale(s->stop_duration_opt, inlink->sample_rate,
AV_TIME_BASE);
s->stop_silence = av_rescale(s->stop_silence_opt, inlink->sample_rate,
AV_TIME_BASE);
s->start_found_periods = 0;
s->stop_found_periods = 0;
return 0;
}
static int config_output(AVFilterLink *outlink)
{
AVFilterContext *ctx = outlink->src;
SilenceRemoveContext *s = ctx->priv;
switch (s->detection) {
case D_AVG:
case D_RMS:
s->cache_size = 1;
break;
case D_DEV:
s->cache_size = 2;
break;
case D_MEDIAN:
case D_PEAK:
case D_PTP:
s->cache_size = s->window_duration;
break;
}
s->start_window = ff_get_audio_buffer(outlink, s->window_duration);
s->stop_window = ff_get_audio_buffer(outlink, s->window_duration);
s->start_cache = av_calloc(outlink->ch_layout.nb_channels, s->cache_size * sizeof(*s->start_cache));
s->stop_cache = av_calloc(outlink->ch_layout.nb_channels, s->cache_size * sizeof(*s->stop_cache));
if (!s->start_window || !s->stop_window || !s->start_cache || !s->stop_cache)
return AVERROR(ENOMEM);
s->start_queuef = ff_get_audio_buffer(outlink, s->start_silence + 1);
s->stop_queuef = ff_get_audio_buffer(outlink, s->stop_silence + 1);
if (!s->start_queuef || !s->stop_queuef)
return AVERROR(ENOMEM);
s->start_front = av_calloc(outlink->ch_layout.nb_channels, sizeof(*s->start_front));
s->start_back = av_calloc(outlink->ch_layout.nb_channels, sizeof(*s->start_back));
s->stop_front = av_calloc(outlink->ch_layout.nb_channels, sizeof(*s->stop_front));
s->stop_back = av_calloc(outlink->ch_layout.nb_channels, sizeof(*s->stop_back));
if (!s->start_front || !s->start_back || !s->stop_front || !s->stop_back)
return AVERROR(ENOMEM);
clear_windows(s);
switch (s->detection) {
case D_AVG:
s->compute_flt = compute_avg_flt;
s->compute_dbl = compute_avg_dbl;
break;
case D_DEV:
s->compute_flt = compute_dev_flt;
s->compute_dbl = compute_dev_dbl;
break;
case D_PTP:
s->compute_flt = compute_ptp_flt;
s->compute_dbl = compute_ptp_dbl;
break;
case D_MEDIAN:
s->compute_flt = compute_median_flt;
s->compute_dbl = compute_median_dbl;
break;
case D_PEAK:
s->compute_flt = compute_peak_flt;
s->compute_dbl = compute_peak_dbl;
break;
case D_RMS:
s->compute_flt = compute_rms_flt;
s->compute_dbl = compute_rms_dbl;
break;
}
return 0;
}
static int filter_frame(AVFilterLink *outlink, AVFrame *in)
{
const int nb_channels = outlink->ch_layout.nb_channels;
AVFilterContext *ctx = outlink->src;
SilenceRemoveContext *s = ctx->priv;
int max_out_nb_samples;
int out_nb_samples = 0;
int in_nb_samples;
const double *srcd;
const float *srcf;
AVFrame *out;
double *dstd;
float *dstf;
if (s->next_pts == AV_NOPTS_VALUE)
s->next_pts = in->pts;
in_nb_samples = in->nb_samples;
max_out_nb_samples = in->nb_samples +
s->start_silence +
s->stop_silence;
if (max_out_nb_samples <= 0) {
av_frame_free(&in);
ff_filter_set_ready(ctx, 100);
return 0;
}
out = ff_get_audio_buffer(outlink, max_out_nb_samples);
if (!out) {
av_frame_free(&in);
return AVERROR(ENOMEM);
}
if (s->timestamp_mode == TS_WRITE)
out->pts = s->next_pts;
else
out->pts = in->pts;
switch (outlink->format) {
case AV_SAMPLE_FMT_FLT:
srcf = (const float *)in->data[0];
dstf = (float *)out->data[0];
if (s->start_periods > 0 && s->stop_periods > 0) {
const float *src = srcf;
if (s->start_found_periods >= 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_start_flt(ctx, src + n * nb_channels,
dstf, &out_nb_samples,
nb_channels);
}
in_nb_samples = out_nb_samples;
out_nb_samples = 0;
src = dstf;
}
for (int n = 0; n < in_nb_samples; n++) {
filter_stop_flt(ctx, src + n * nb_channels,
dstf, &out_nb_samples,
nb_channels);
}
} else if (s->start_periods > 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_start_flt(ctx, srcf + n * nb_channels,
dstf, &out_nb_samples,
nb_channels);
}
} else if (s->stop_periods > 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_stop_flt(ctx, srcf + n * nb_channels,
dstf, &out_nb_samples,
nb_channels);
}
}
break;
case AV_SAMPLE_FMT_DBL:
srcd = (const double *)in->data[0];
dstd = (double *)out->data[0];
if (s->start_periods > 0 && s->stop_periods > 0) {
const double *src = srcd;
if (s->start_found_periods >= 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_start_dbl(ctx, src + n * nb_channels,
dstd, &out_nb_samples,
nb_channels);
}
in_nb_samples = out_nb_samples;
out_nb_samples = 0;
src = dstd;
}
for (int n = 0; n < in_nb_samples; n++) {
filter_stop_dbl(ctx, src + n * nb_channels,
dstd, &out_nb_samples,
nb_channels);
}
} else if (s->start_periods > 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_start_dbl(ctx, srcd + n * nb_channels,
dstd, &out_nb_samples,
nb_channels);
}
} else if (s->stop_periods > 0) {
for (int n = 0; n < in_nb_samples; n++) {
filter_stop_dbl(ctx, srcd + n * nb_channels,
dstd, &out_nb_samples,
nb_channels);
}
}
break;
}
av_frame_free(&in);
if (out_nb_samples > 0) {
s->next_pts += out_nb_samples;
out->nb_samples = out_nb_samples;
return ff_filter_frame(outlink, out);
}
av_frame_free(&out);
ff_filter_set_ready(ctx, 100);
return 0;
}
static int activate(AVFilterContext *ctx)
{
AVFilterLink *outlink = ctx->outputs[0];
AVFilterLink *inlink = ctx->inputs[0];
SilenceRemoveContext *s = ctx->priv;
AVFrame *in;
int ret;
FF_FILTER_FORWARD_STATUS_BACK(outlink, inlink);
ret = ff_inlink_consume_frame(inlink, &in);
if (ret < 0)
return ret;
if (ret > 0) {
if (s->start_periods == 1 && s->stop_periods == 0 &&
s->start_found_periods < 0) {
if (s->timestamp_mode == TS_WRITE)
in->pts = s->next_pts;
s->next_pts += in->nb_samples;
return ff_filter_frame(outlink, in);
}
if (s->start_periods == 0 && s->stop_periods == 0)
return ff_filter_frame(outlink, in);
return filter_frame(outlink, in);
}
FF_FILTER_FORWARD_STATUS(inlink, outlink);
FF_FILTER_FORWARD_WANTED(outlink, inlink);
return FFERROR_NOT_READY;
}
static av_cold void uninit(AVFilterContext *ctx)
{
SilenceRemoveContext *s = ctx->priv;
av_frame_free(&s->start_window);
av_frame_free(&s->stop_window);
av_frame_free(&s->start_queuef);
av_frame_free(&s->stop_queuef);
av_freep(&s->start_cache);
av_freep(&s->stop_cache);
av_freep(&s->start_front);
av_freep(&s->start_back);
av_freep(&s->stop_front);
av_freep(&s->stop_back);
}
static const AVFilterPad silenceremove_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_AUDIO,
.config_props = config_input,
},
};
static const AVFilterPad silenceremove_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_AUDIO,
.config_props = config_output,
},
};
const AVFilter ff_af_silenceremove = {
.name = "silenceremove",
.description = NULL_IF_CONFIG_SMALL("Remove silence."),
.priv_size = sizeof(SilenceRemoveContext),
.priv_class = &silenceremove_class,
.init = init,
.activate = activate,
.uninit = uninit,
FILTER_INPUTS(silenceremove_inputs),
FILTER_OUTPUTS(silenceremove_outputs),
FILTER_SAMPLEFMTS(AV_SAMPLE_FMT_FLT,
AV_SAMPLE_FMT_DBL),
.process_command = ff_filter_process_command,
.flags = AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL,
};