1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-28 20:53:54 +02:00
FFmpeg/libavfilter/vf_telecine.c
Anton Khirnov 1e7d2007c3 all: use designated initializers for AVOption.unit
Makes it robust against adding fields before it, which will be useful in
following commits.

Majority of the patch generated by the following Coccinelle script:

@@
typedef AVOption;
identifier arr_name;
initializer list il;
initializer list[8] il1;
expression tail;
@@
AVOption arr_name[] = { il, { il1,
- tail
+ .unit = tail
}, ...  };

with some manual changes, as the script:
* has trouble with options defined inside macros
* sometimes does not handle options under an #else branch
* sometimes swallows whitespace
2024-02-14 14:53:41 +01:00

329 lines
11 KiB
C

/*
* Copyright (c) 2012 Rudolf Polzer
* Copyright (c) 2013 Paul B Mahol
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file telecine filter, heavily based from mpv-player:TOOLS/vf_dlopen/telecine.c by
* Rudolf Polzer.
*/
#include "libavutil/avstring.h"
#include "libavutil/imgutils.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "avfilter.h"
#include "filters.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
typedef struct TelecineContext {
const AVClass *class;
int first_field;
char *pattern;
unsigned int pattern_pos;
int64_t start_time;
AVRational pts;
AVRational ts_unit;
int out_cnt;
int occupied;
int nb_planes;
int planeheight[4];
int stride[4];
AVFrame *frame[5];
AVFrame *temp;
} TelecineContext;
#define OFFSET(x) offsetof(TelecineContext, x)
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
static const AVOption telecine_options[] = {
{"first_field", "select first field", OFFSET(first_field), AV_OPT_TYPE_INT, {.i64=0}, 0, 1, FLAGS, .unit = "field"},
{"top", "select top field first", 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, .unit = "field"},
{"t", "select top field first", 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, .unit = "field"},
{"bottom", "select bottom field first", 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, .unit = "field"},
{"b", "select bottom field first", 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, .unit = "field"},
{"pattern", "pattern that describe for how many fields a frame is to be displayed", OFFSET(pattern), AV_OPT_TYPE_STRING, {.str="23"}, 0, 0, FLAGS},
{NULL}
};
AVFILTER_DEFINE_CLASS(telecine);
static av_cold int init(AVFilterContext *ctx)
{
TelecineContext *s = ctx->priv;
const char *p;
int max = 0;
if (!strlen(s->pattern)) {
av_log(ctx, AV_LOG_ERROR, "No pattern provided.\n");
return AVERROR_INVALIDDATA;
}
for (p = s->pattern; *p; p++) {
if (!av_isdigit(*p)) {
av_log(ctx, AV_LOG_ERROR, "Provided pattern includes non-numeric characters.\n");
return AVERROR_INVALIDDATA;
}
max = FFMAX(*p - '0', max);
s->pts.num += 2;
s->pts.den += *p - '0';
}
s->start_time = AV_NOPTS_VALUE;
s->out_cnt = (max + 1) / 2;
av_log(ctx, AV_LOG_INFO, "Telecine pattern %s yields up to %d frames per frame, pts advance factor: %d/%d\n",
s->pattern, s->out_cnt, s->pts.num, s->pts.den);
return 0;
}
static int query_formats(AVFilterContext *ctx)
{
int reject_flags = AV_PIX_FMT_FLAG_BITSTREAM |
AV_PIX_FMT_FLAG_HWACCEL |
AV_PIX_FMT_FLAG_PAL;
return ff_set_common_formats(ctx, ff_formats_pixdesc_filter(0, reject_flags));
}
static int config_input(AVFilterLink *inlink)
{
TelecineContext *s = inlink->dst->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
int i, ret;
s->temp = ff_get_video_buffer(inlink, inlink->w, inlink->h);
if (!s->temp)
return AVERROR(ENOMEM);
for (i = 0; i < s->out_cnt; i++) {
s->frame[i] = ff_get_video_buffer(inlink, inlink->w, inlink->h);
if (!s->frame[i])
return AVERROR(ENOMEM);
}
if ((ret = av_image_fill_linesizes(s->stride, inlink->format, inlink->w)) < 0)
return ret;
s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
s->planeheight[0] = s->planeheight[3] = inlink->h;
s->nb_planes = av_pix_fmt_count_planes(inlink->format);
return 0;
}
static int config_output(AVFilterLink *outlink)
{
AVFilterContext *ctx = outlink->src;
TelecineContext *s = ctx->priv;
const AVFilterLink *inlink = ctx->inputs[0];
AVRational fps = inlink->frame_rate;
if (!fps.num || !fps.den) {
av_log(ctx, AV_LOG_ERROR, "The input needs a constant frame rate; "
"current rate of %d/%d is invalid\n", fps.num, fps.den);
return AVERROR(EINVAL);
}
fps = av_mul_q(fps, av_inv_q(s->pts));
av_log(ctx, AV_LOG_VERBOSE, "FPS: %d/%d -> %d/%d\n",
inlink->frame_rate.num, inlink->frame_rate.den, fps.num, fps.den);
outlink->frame_rate = fps;
outlink->time_base = av_mul_q(inlink->time_base, s->pts);
av_log(ctx, AV_LOG_VERBOSE, "TB: %d/%d -> %d/%d\n",
inlink->time_base.num, inlink->time_base.den, outlink->time_base.num, outlink->time_base.den);
s->ts_unit = av_inv_q(av_mul_q(fps, outlink->time_base));
return 0;
}
static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
{
AVFilterContext *ctx = inlink->dst;
AVFilterLink *outlink = ctx->outputs[0];
TelecineContext *s = ctx->priv;
int i, len, ret = 0, nout = 0;
if (s->start_time == AV_NOPTS_VALUE)
s->start_time = inpicref->pts;
len = s->pattern[s->pattern_pos] - '0';
s->pattern_pos++;
if (!s->pattern[s->pattern_pos])
s->pattern_pos = 0;
if (!len) { // do not output any field from this frame
av_frame_free(&inpicref);
return 0;
}
if (s->occupied) {
ret = ff_inlink_make_frame_writable(inlink, &s->frame[nout]);
if (ret < 0) {
av_frame_free(&inpicref);
return ret;
}
for (i = 0; i < s->nb_planes; i++) {
// fill in the EARLIER field from the buffered pic
av_image_copy_plane(s->frame[nout]->data[i] + s->frame[nout]->linesize[i] * s->first_field,
s->frame[nout]->linesize[i] * 2,
s->temp->data[i] + s->temp->linesize[i] * s->first_field,
s->temp->linesize[i] * 2,
s->stride[i],
(s->planeheight[i] - s->first_field + 1) / 2);
// fill in the LATER field from the new pic
av_image_copy_plane(s->frame[nout]->data[i] + s->frame[nout]->linesize[i] * !s->first_field,
s->frame[nout]->linesize[i] * 2,
inpicref->data[i] + inpicref->linesize[i] * !s->first_field,
inpicref->linesize[i] * 2,
s->stride[i],
(s->planeheight[i] - !s->first_field + 1) / 2);
}
#if FF_API_INTERLACED_FRAME
FF_DISABLE_DEPRECATION_WARNINGS
s->frame[nout]->interlaced_frame = 1;
s->frame[nout]->top_field_first = !s->first_field;
FF_ENABLE_DEPRECATION_WARNINGS
#endif
s->frame[nout]->flags |= AV_FRAME_FLAG_INTERLACED;
if (s->first_field)
s->frame[nout]->flags &= ~AV_FRAME_FLAG_TOP_FIELD_FIRST;
else
s->frame[nout]->flags |= AV_FRAME_FLAG_TOP_FIELD_FIRST;
nout++;
len--;
s->occupied = 0;
}
while (len >= 2) {
// output THIS image as-is
ret = ff_inlink_make_frame_writable(inlink, &s->frame[nout]);
if (ret < 0) {
av_frame_free(&inpicref);
return ret;
}
for (i = 0; i < s->nb_planes; i++)
av_image_copy_plane(s->frame[nout]->data[i], s->frame[nout]->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
#if FF_API_INTERLACED_FRAME
FF_DISABLE_DEPRECATION_WARNINGS
s->frame[nout]->interlaced_frame = inpicref->interlaced_frame;
s->frame[nout]->top_field_first = inpicref->top_field_first;
FF_ENABLE_DEPRECATION_WARNINGS
#endif
s->frame[nout]->flags |= (inpicref->flags & (AV_FRAME_FLAG_INTERLACED | AV_FRAME_FLAG_TOP_FIELD_FIRST));
nout++;
len -= 2;
}
if (len >= 1) {
// copy THIS image to the buffer, we need it later
for (i = 0; i < s->nb_planes; i++)
av_image_copy_plane(s->temp->data[i], s->temp->linesize[i],
inpicref->data[i], inpicref->linesize[i],
s->stride[i],
s->planeheight[i]);
s->occupied = 1;
}
for (i = 0; i < nout; i++) {
AVFrame *frame = av_frame_clone(s->frame[i]);
int interlaced = frame ? !!(frame->flags & AV_FRAME_FLAG_INTERLACED) : 0;
int tff = frame ? !!(frame->flags & AV_FRAME_FLAG_TOP_FIELD_FIRST) : 0;
if (!frame) {
av_frame_free(&inpicref);
return AVERROR(ENOMEM);
}
av_frame_copy_props(frame, inpicref);
#if FF_API_INTERLACED_FRAME
FF_DISABLE_DEPRECATION_WARNINGS
frame->interlaced_frame = interlaced;
frame->top_field_first = tff;
FF_ENABLE_DEPRECATION_WARNINGS
#endif
if (interlaced)
frame->flags |= AV_FRAME_FLAG_INTERLACED;
else
frame->flags &= ~AV_FRAME_FLAG_INTERLACED;
if (tff)
frame->flags |= AV_FRAME_FLAG_TOP_FIELD_FIRST;
else
frame->flags &= ~AV_FRAME_FLAG_TOP_FIELD_FIRST;
frame->pts = ((s->start_time == AV_NOPTS_VALUE) ? 0 : s->start_time) +
av_rescale(outlink->frame_count_in, s->ts_unit.num,
s->ts_unit.den);
ret = ff_filter_frame(outlink, frame);
}
av_frame_free(&inpicref);
return ret;
}
static av_cold void uninit(AVFilterContext *ctx)
{
TelecineContext *s = ctx->priv;
int i;
av_frame_free(&s->temp);
for (i = 0; i < s->out_cnt; i++)
av_frame_free(&s->frame[i]);
}
static const AVFilterPad telecine_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = filter_frame,
.config_props = config_input,
},
};
static const AVFilterPad telecine_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_output,
},
};
const AVFilter ff_vf_telecine = {
.name = "telecine",
.description = NULL_IF_CONFIG_SMALL("Apply a telecine pattern."),
.priv_size = sizeof(TelecineContext),
.priv_class = &telecine_class,
.init = init,
.uninit = uninit,
FILTER_INPUTS(telecine_inputs),
FILTER_OUTPUTS(telecine_outputs),
FILTER_QUERY_FUNC(query_formats),
};