1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-23 12:43:46 +02:00

Add lensfun filter

Lensfun is a library that applies lens correction to an image using a
database of cameras/lenses (you provide the camera and lens models, and
it uses the corresponding database entry's parameters to apply lens
correction). It is licensed under LGPL3.

The lensfun filter utilizes the lensfun library to apply lens
correction to videos as well as images.

This filter was created out of necessity since I wanted to apply lens
correction to a video and the lenscorrection filter did not work for me.

While this filter requires little info from the user to apply lens
correction, the flaw is that lensfun is intended to be used on indvidual
images. When used on a video, the parameters such as focal length is
constant, so lens correction may fail on videos where the camera's focal
length changes (zooming in or out via zoom lens). To use this filter
correctly on videos where such parameters change, timeline editing may
be used since this filter supports it.

Note that valgrind shows a small memory leak which is not from this
filter but from the lensfun library (memory is allocated when loading
the lensfun database but it somehow isn't deallocated even during
cleanup; it is briefly created in the init function of the filter, and
destroyed before the init function returns). This may have been fixed by
the latest commit in the lensfun repository; the current latest release
of lensfun is almost 3 years ago.

Bi-Linear interpolation is used by default as lanczos interpolation
shows more artifacts in the corrected image in my tests.

The lanczos interpolation is derived from lenstool's implementation of
lanczos interpolation. Lenstool is an app within the lensfun repository
which is licensed under GPL3.

v2 of this patch fixes license notice in libavfilter/vf_lensfun.c

v3 of this patch fixes code style and dependency to gplv3 (thanks to
Paul B Mahol for pointing out the mentioned issues).

v4 of this patch fixes more code style issues that were missed in
v3.

v5 of this patch adds line breaks to some of the documentation in
doc/filters.texi (thanks to Gyan Doshi for pointing out the issue).

v6 of this patch fixes more problems (thanks to Moritz Barsnick for
pointing them out).

v7 of this patch fixes use of sqrt() (changed to sqrtf(); thanks to
Moritz Barsnick for pointing this out). Also should be rebased off of
latest master branch commits at this point.

Signed-off-by: Stephen Seo <seo.disparate@gmail.com>
This commit is contained in:
Stephen Seo 2018-07-13 19:33:12 +09:00 committed by Paul B Mahol
parent c5329d64b1
commit 0ea20124b7
5 changed files with 672 additions and 0 deletions

4
configure vendored
View File

@ -239,6 +239,7 @@ External library support:
--enable-libilbc enable iLBC de/encoding via libilbc [no]
--enable-libjack enable JACK audio sound server [no]
--enable-libkvazaar enable HEVC encoding via libkvazaar [no]
--enable-liblensfun enable lensfun lens correction [no]
--enable-libmodplug enable ModPlug via libmodplug [no]
--enable-libmp3lame enable MP3 encoding via libmp3lame [no]
--enable-libopencore-amrnb enable AMR-NB de/encoding via libopencore-amrnb [no]
@ -1656,6 +1657,7 @@ EXTERNAL_LIBRARY_NONFREE_LIST="
EXTERNAL_LIBRARY_VERSION3_LIST="
gmp
liblensfun
libopencore_amrnb
libopencore_amrwb
libvmaf
@ -3353,6 +3355,7 @@ hqdn3d_filter_deps="gpl"
interlace_filter_deps="gpl"
kerndeint_filter_deps="gpl"
ladspa_filter_deps="ladspa libdl"
lensfun_filter_deps="liblensfun version3"
lv2_filter_deps="lv2"
mcdeint_filter_deps="avcodec gpl"
movie_filter_deps="avcodec avformat"
@ -6023,6 +6026,7 @@ enabled libgsm && { for gsm_hdr in "gsm.h" "gsm/gsm.h"; do
done || die "ERROR: libgsm not found"; }
enabled libilbc && require libilbc ilbc.h WebRtcIlbcfix_InitDecode -lilbc $pthreads_extralibs
enabled libkvazaar && require_pkg_config libkvazaar "kvazaar >= 0.8.1" kvazaar.h kvz_api_get
enabled liblensfun && require_pkg_config liblensfun lensfun lensfun.h lf_db_new
# While it may appear that require is being used as a pkg-config
# fallback for libmfx, it is actually being used to detect a different
# installation route altogether. If libmfx is installed via the Intel

View File

@ -10700,6 +10700,124 @@ The formula that generates the correction is:
where @var{r_0} is halve of the image diagonal and @var{r_src} and @var{r_tgt} are the
distances from the focal point in the source and target images, respectively.
@section lensfun
Apply lens correction via the lensfun library (@url{http://lensfun.sourceforge.net/}).
The @code{lensfun} filter requires the camera make, camera model, and lens model
to apply the lens correction. The filter will load the lensfun database and
query it to find the corresponding camera and lens entries in the database. As
long as these entries can be found with the given options, the filter can
perform corrections on frames. Note that incomplete strings will result in the
filter choosing the best match with the given options, and the filter will
output the chosen camera and lens models (logged with level "info"). You must
provide the make, camera model, and lens model as they are required.
The filter accepts the following options:
@table @option
@item make
The make of the camera (for example, "Canon"). This option is required.
@item model
The model of the camera (for example, "Canon EOS 100D"). This option is
required.
@item lens_model
The model of the lens (for example, "Canon EF-S 18-55mm f/3.5-5.6 IS STM"). This
option is required.
@item mode
The type of correction to apply. The following values are valid options:
@table @samp
@item vignetting
Enables fixing lens vignetting.
@item geometry
Enables fixing lens geometry. This is the default.
@item subpixel
Enables fixing chromatic aberrations.
@item vig_geo
Enables fixing lens vignetting and lens geometry.
@item vig_subpixel
Enables fixing lens vignetting and chromatic aberrations.
@item distortion
Enables fixing both lens geometry and chromatic aberrations.
@item all
Enables all possible corrections.
@end table
@item focal_length
The focal length of the image/video (zoom; expected constant for video). For
example, a 18--55mm lens has focal length range of [18--55], so a value in that
range should be chosen when using that lens. Default 18.
@item aperture
The aperture of the image/video (expected constant for video). Note that
aperture is only used for vignetting correction. Default 3.5.
@item focus_distance
The focus distance of the image/video (expected constant for video). Note that
focus distance is only used for vignetting and only slightly affects the
vignetting correction process. If unknown, leave it at the default value (which
is 1000).
@item target_geometry
The target geometry of the output image/video. The following values are valid
options:
@table @samp
@item rectilinear (default)
@item fisheye
@item panoramic
@item equirectangular
@item fisheye_orthographic
@item fisheye_stereographic
@item fisheye_equisolid
@item fisheye_thoby
@end table
@item reverse
Apply the reverse of image correction (instead of correcting distortion, apply
it).
@item interpolation
The type of interpolation used when correcting distortion. The following values
are valid options:
@table @samp
@item nearest
@item linear (default)
@item lanczos
@end table
@end table
@subsection Examples
@itemize
@item
Apply lens correction with make "Canon", camera model "Canon EOS 100D", and lens
model "Canon EF-S 18-55mm f/3.5-5.6 IS STM" with focal length of "18" and
aperture of "8.0".
@example
ffmpeg -i input.mov -vf lensfun=make=Canon:model="Canon EOS 100D":lens_model="Canon EF-S 18-55mm f/3.5-5.6 IS STM":focal_length=18:aperture=8 -c:v h264 -b:v 8000k output.mov
@end example
@item
Apply the same as before, but only for the first 5 seconds of video.
@example
ffmpeg -i input.mov -vf lensfun=make=Canon:model="Canon EOS 100D":lens_model="Canon EF-S 18-55mm f/3.5-5.6 IS STM":focal_length=18:aperture=8:enable='lte(t\,5)' -c:v h264 -b:v 8000k output.mov
@end example
@end itemize
@section libvmaf
Obtain the VMAF (Video Multi-Method Assessment Fusion)

View File

@ -248,6 +248,7 @@ OBJS-$(CONFIG_INTERLACE_FILTER) += vf_tinterlace.o
OBJS-$(CONFIG_INTERLEAVE_FILTER) += f_interleave.o
OBJS-$(CONFIG_KERNDEINT_FILTER) += vf_kerndeint.o
OBJS-$(CONFIG_LENSCORRECTION_FILTER) += vf_lenscorrection.o
OBJS-$(CONFIG_LENSFUN_FILTER) += vf_lensfun.o
OBJS-$(CONFIG_LIBVMAF_FILTER) += vf_libvmaf.o framesync.o
OBJS-$(CONFIG_LIMITER_FILTER) += vf_limiter.o
OBJS-$(CONFIG_LOOP_FILTER) += f_loop.o

View File

@ -237,6 +237,7 @@ extern AVFilter ff_vf_interlace;
extern AVFilter ff_vf_interleave;
extern AVFilter ff_vf_kerndeint;
extern AVFilter ff_vf_lenscorrection;
extern AVFilter ff_vf_lensfun;
extern AVFilter ff_vf_libvmaf;
extern AVFilter ff_vf_limiter;
extern AVFilter ff_vf_loop;

548
libavfilter/vf_lensfun.c Normal file
View File

@ -0,0 +1,548 @@
/*
* Copyright (C) 2007 by Andrew Zabolotny (author of lensfun, from which this filter derives from)
* Copyright (C) 2018 Stephen Seo
*
* This file is part of FFmpeg.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see <https://www.gnu.org/licenses/>.
*/
/**
* @file
* Lensfun filter, applies lens correction with parameters from the lensfun database
*
* @see https://lensfun.sourceforge.net/
*/
#include <float.h>
#include <math.h>
#include "libavutil/avassert.h"
#include "libavutil/imgutils.h"
#include "libavutil/opt.h"
#include "libswscale/swscale.h"
#include "avfilter.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
#include <lensfun.h>
#define LANCZOS_RESOLUTION 256
enum Mode {
VIGNETTING = 0x1,
GEOMETRY_DISTORTION = 0x2,
SUBPIXEL_DISTORTION = 0x4
};
enum InterpolationType {
NEAREST,
LINEAR,
LANCZOS
};
typedef struct VignettingThreadData {
int width, height;
uint8_t *data_in;
int linesize_in;
int pixel_composition;
lfModifier *modifier;
} VignettingThreadData;
typedef struct DistortionCorrectionThreadData {
int width, height;
const float *distortion_coords;
const uint8_t *data_in;
uint8_t *data_out;
int linesize_in, linesize_out;
const float *interpolation;
int mode;
int interpolation_type;
} DistortionCorrectionThreadData;
typedef struct LensfunContext {
const AVClass *class;
const char *make, *model, *lens_model;
int mode;
float focal_length;
float aperture;
float focus_distance;
int target_geometry;
int reverse;
int interpolation_type;
float *distortion_coords;
float *interpolation;
lfLens *lens;
lfCamera *camera;
lfModifier *modifier;
} LensfunContext;
#define OFFSET(x) offsetof(LensfunContext, x)
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
static const AVOption lensfun_options[] = {
{ "make", "set camera maker", OFFSET(make), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, FLAGS },
{ "model", "set camera model", OFFSET(model), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, FLAGS },
{ "lens_model", "set lens model", OFFSET(lens_model), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, FLAGS },
{ "mode", "set mode", OFFSET(mode), AV_OPT_TYPE_INT, {.i64=GEOMETRY_DISTORTION}, 0, VIGNETTING | GEOMETRY_DISTORTION | SUBPIXEL_DISTORTION, FLAGS, "mode" },
{ "vignetting", "fix lens vignetting", 0, AV_OPT_TYPE_CONST, {.i64=VIGNETTING}, 0, 0, FLAGS, "mode" },
{ "geometry", "correct geometry distortion", 0, AV_OPT_TYPE_CONST, {.i64=GEOMETRY_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "subpixel", "fix chromatic aberrations", 0, AV_OPT_TYPE_CONST, {.i64=SUBPIXEL_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "vig_geo", "fix lens vignetting and correct geometry distortion", 0, AV_OPT_TYPE_CONST, {.i64=VIGNETTING | GEOMETRY_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "vig_subpixel", "fix lens vignetting and chromatic aberrations", 0, AV_OPT_TYPE_CONST, {.i64=VIGNETTING | SUBPIXEL_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "distortion", "correct geometry distortion and chromatic aberrations", 0, AV_OPT_TYPE_CONST, {.i64=GEOMETRY_DISTORTION | SUBPIXEL_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "all", NULL, 0, AV_OPT_TYPE_CONST, {.i64=VIGNETTING | GEOMETRY_DISTORTION | SUBPIXEL_DISTORTION}, 0, 0, FLAGS, "mode" },
{ "focal_length", "focal length of video (zoom; constant for the duration of the use of this filter)", OFFSET(focal_length), AV_OPT_TYPE_FLOAT, {.dbl=18}, 0.0, DBL_MAX, FLAGS },
{ "aperture", "aperture (constant for the duration of the use of this filter)", OFFSET(aperture), AV_OPT_TYPE_FLOAT, {.dbl=3.5}, 0.0, DBL_MAX, FLAGS },
{ "focus_distance", "focus distance (constant for the duration of the use of this filter)", OFFSET(focus_distance), AV_OPT_TYPE_FLOAT, {.dbl=1000.0f}, 0.0, DBL_MAX, FLAGS },
{ "target_geometry", "target geometry of the lens correction (only when geometry correction is enabled)", OFFSET(target_geometry), AV_OPT_TYPE_INT, {.i64=LF_RECTILINEAR}, 0, INT_MAX, FLAGS, "lens_geometry" },
{ "rectilinear", "rectilinear lens (default)", 0, AV_OPT_TYPE_CONST, {.i64=LF_RECTILINEAR}, 0, 0, FLAGS, "lens_geometry" },
{ "fisheye", "fisheye lens", 0, AV_OPT_TYPE_CONST, {.i64=LF_FISHEYE}, 0, 0, FLAGS, "lens_geometry" },
{ "panoramic", "panoramic (cylindrical)", 0, AV_OPT_TYPE_CONST, {.i64=LF_PANORAMIC}, 0, 0, FLAGS, "lens_geometry" },
{ "equirectangular", "equirectangular", 0, AV_OPT_TYPE_CONST, {.i64=LF_EQUIRECTANGULAR}, 0, 0, FLAGS, "lens_geometry" },
{ "fisheye_orthographic", "orthographic fisheye", 0, AV_OPT_TYPE_CONST, {.i64=LF_FISHEYE_ORTHOGRAPHIC}, 0, 0, FLAGS, "lens_geometry" },
{ "fisheye_stereographic", "stereographic fisheye", 0, AV_OPT_TYPE_CONST, {.i64=LF_FISHEYE_STEREOGRAPHIC}, 0, 0, FLAGS, "lens_geometry" },
{ "fisheye_equisolid", "equisolid fisheye", 0, AV_OPT_TYPE_CONST, {.i64=LF_FISHEYE_EQUISOLID}, 0, 0, FLAGS, "lens_geometry" },
{ "fisheye_thoby", "fisheye as measured by thoby", 0, AV_OPT_TYPE_CONST, {.i64=LF_FISHEYE_THOBY}, 0, 0, FLAGS, "lens_geometry" },
{ "reverse", "Does reverse correction (regular image to lens distorted)", OFFSET(reverse), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS },
{ "interpolation", "Type of interpolation", OFFSET(interpolation_type), AV_OPT_TYPE_INT, {.i64=LINEAR}, 0, LANCZOS, FLAGS, "interpolation" },
{ "nearest", NULL, 0, AV_OPT_TYPE_CONST, {.i64=NEAREST}, 0, 0, FLAGS, "interpolation" },
{ "linear", NULL, 0, AV_OPT_TYPE_CONST, {.i64=LINEAR}, 0, 0, FLAGS, "interpolation" },
{ "lanczos", NULL, 0, AV_OPT_TYPE_CONST, {.i64=LANCZOS}, 0, 0, FLAGS, "interpolation" },
{ NULL }
};
AVFILTER_DEFINE_CLASS(lensfun);
static av_cold int init(AVFilterContext *ctx)
{
LensfunContext *lensfun = ctx->priv;
lfDatabase *db;
const lfCamera **cameras;
const lfLens **lenses;
if (!lensfun->make) {
av_log(ctx, AV_LOG_FATAL, "Option \"make\" not specified\n");
return AVERROR(EINVAL);
} else if (!lensfun->model) {
av_log(ctx, AV_LOG_FATAL, "Option \"model\" not specified\n");
return AVERROR(EINVAL);
} else if (!lensfun->lens_model) {
av_log(ctx, AV_LOG_FATAL, "Option \"lens_model\" not specified\n");
return AVERROR(EINVAL);
}
lensfun->lens = lf_lens_new();
lensfun->camera = lf_camera_new();
db = lf_db_new();
if (lf_db_load(db) != LF_NO_ERROR) {
lf_db_destroy(db);
av_log(ctx, AV_LOG_FATAL, "Failed to load lensfun database\n");
return AVERROR_INVALIDDATA;
}
cameras = lf_db_find_cameras(db, lensfun->make, lensfun->model);
if (cameras && *cameras) {
lf_camera_copy(lensfun->camera, *cameras);
av_log(ctx, AV_LOG_INFO, "Using camera %s\n", lensfun->camera->Model);
} else {
lf_free(cameras);
lf_db_destroy(db);
av_log(ctx, AV_LOG_FATAL, "Failed to find camera in lensfun database\n");
return AVERROR_INVALIDDATA;
}
lf_free(cameras);
lenses = lf_db_find_lenses_hd(db, lensfun->camera, NULL, lensfun->lens_model, 0);
if (lenses && *lenses) {
lf_lens_copy(lensfun->lens, *lenses);
av_log(ctx, AV_LOG_INFO, "Using lens %s\n", lensfun->lens->Model);
} else {
lf_free(lenses);
lf_db_destroy(db);
av_log(ctx, AV_LOG_FATAL, "Failed to find lens in lensfun database\n");
return AVERROR_INVALIDDATA;
}
lf_free(lenses);
lf_db_destroy(db);
return 0;
}
static int query_formats(AVFilterContext *ctx)
{
// Some of the functions provided by lensfun require pixels in RGB format
static const enum AVPixelFormat fmts[] = {AV_PIX_FMT_RGB24, AV_PIX_FMT_NONE};
AVFilterFormats *fmts_list = ff_make_format_list(fmts);
return ff_set_common_formats(ctx, fmts_list);
}
static float lanczos_kernel(float x)
{
if (x == 0.0f) {
return 1.0f;
} else if (x > -2.0f && x < 2.0f) {
return (2.0f * sin(M_PI * x) * sin(M_PI / 2.0f * x)) / (M_PI * M_PI * x * x);
} else {
return 0.0f;
}
}
static int config_props(AVFilterLink *inlink)
{
AVFilterContext *ctx = inlink->dst;
LensfunContext *lensfun = ctx->priv;
int index;
float a;
int lensfun_mode = 0;
if (!lensfun->modifier) {
if (lensfun->camera && lensfun->lens) {
lensfun->modifier = lf_modifier_new(lensfun->lens,
lensfun->camera->CropFactor,
inlink->w,
inlink->h);
if (lensfun->mode & VIGNETTING)
lensfun_mode |= LF_MODIFY_VIGNETTING;
if (lensfun->mode & GEOMETRY_DISTORTION)
lensfun_mode |= LF_MODIFY_DISTORTION | LF_MODIFY_GEOMETRY | LF_MODIFY_SCALE;
if (lensfun->mode & SUBPIXEL_DISTORTION)
lensfun_mode |= LF_MODIFY_TCA;
lf_modifier_initialize(lensfun->modifier,
lensfun->lens,
LF_PF_U8,
lensfun->focal_length,
lensfun->aperture,
lensfun->focus_distance,
0.0,
lensfun->target_geometry,
lensfun_mode,
lensfun->reverse);
} else {
// lensfun->camera and lensfun->lens should have been initialized
return AVERROR_BUG;
}
}
if (!lensfun->distortion_coords) {
if (lensfun->mode & SUBPIXEL_DISTORTION) {
lensfun->distortion_coords = av_malloc(sizeof(float) * inlink->w * inlink->h * 2 * 3);
if (!lensfun->distortion_coords)
return AVERROR(ENOMEM);
if (lensfun->mode & GEOMETRY_DISTORTION) {
// apply both geometry and subpixel distortion
lf_modifier_apply_subpixel_geometry_distortion(lensfun->modifier,
0, 0,
inlink->w, inlink->h,
lensfun->distortion_coords);
} else {
// apply only subpixel distortion
lf_modifier_apply_subpixel_distortion(lensfun->modifier,
0, 0,
inlink->w, inlink->h,
lensfun->distortion_coords);
}
} else if (lensfun->mode & GEOMETRY_DISTORTION) {
lensfun->distortion_coords = av_malloc(sizeof(float) * inlink->w * inlink->h * 2);
if (!lensfun->distortion_coords)
return AVERROR(ENOMEM);
// apply only geometry distortion
lf_modifier_apply_geometry_distortion(lensfun->modifier,
0, 0,
inlink->w, inlink->h,
lensfun->distortion_coords);
}
}
if (!lensfun->interpolation)
if (lensfun->interpolation_type == LANCZOS) {
lensfun->interpolation = av_malloc(sizeof(float) * 4 * LANCZOS_RESOLUTION);
if (!lensfun->interpolation)
return AVERROR(ENOMEM);
for (index = 0; index < 4 * LANCZOS_RESOLUTION; ++index) {
if (index == 0) {
lensfun->interpolation[index] = 1.0f;
} else {
a = sqrtf((float)index / LANCZOS_RESOLUTION);
lensfun->interpolation[index] = lanczos_kernel(a);
}
}
}
return 0;
}
static int vignetting_filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
{
const VignettingThreadData *thread_data = arg;
const int slice_start = thread_data->height * jobnr / nb_jobs;
const int slice_end = thread_data->height * (jobnr + 1) / nb_jobs;
lf_modifier_apply_color_modification(thread_data->modifier,
thread_data->data_in + slice_start * thread_data->linesize_in,
0,
slice_start,
thread_data->width,
slice_end - slice_start,
thread_data->pixel_composition,
thread_data->linesize_in);
return 0;
}
static float square(float x)
{
return x * x;
}
static int distortion_correction_filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
{
const DistortionCorrectionThreadData *thread_data = arg;
const int slice_start = thread_data->height * jobnr / nb_jobs;
const int slice_end = thread_data->height * (jobnr + 1) / nb_jobs;
int x, y, i, j, rgb_index;
float interpolated, new_x, new_y, d, norm;
int new_x_int, new_y_int;
for (y = slice_start; y < slice_end; ++y)
for (x = 0; x < thread_data->width; ++x)
for (rgb_index = 0; rgb_index < 3; ++rgb_index) {
if (thread_data->mode & SUBPIXEL_DISTORTION) {
// subpixel (and possibly geometry) distortion correction was applied, correct distortion
switch(thread_data->interpolation_type) {
case NEAREST:
new_x_int = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2] + 0.5f;
new_y_int = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2 + 1] + 0.5f;
if (new_x_int < 0 || new_x_int >= thread_data->width || new_y_int < 0 || new_y_int >= thread_data->height) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = thread_data->data_in[new_x_int * 3 + rgb_index + new_y_int * thread_data->linesize_in];
}
break;
case LINEAR:
interpolated = 0.0f;
new_x = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2];
new_x_int = new_x;
new_y = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2 + 1];
new_y_int = new_y;
if (new_x_int < 0 || new_x_int + 1 >= thread_data->width || new_y_int < 0 || new_y_int + 1 >= thread_data->height) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] =
thread_data->data_in[ new_x_int * 3 + rgb_index + new_y_int * thread_data->linesize_in] * (new_x_int + 1 - new_x) * (new_y_int + 1 - new_y)
+ thread_data->data_in[(new_x_int + 1) * 3 + rgb_index + new_y_int * thread_data->linesize_in] * (new_x - new_x_int) * (new_y_int + 1 - new_y)
+ thread_data->data_in[ new_x_int * 3 + rgb_index + (new_y_int + 1) * thread_data->linesize_in] * (new_x_int + 1 - new_x) * (new_y - new_y_int)
+ thread_data->data_in[(new_x_int + 1) * 3 + rgb_index + (new_y_int + 1) * thread_data->linesize_in] * (new_x - new_x_int) * (new_y - new_y_int);
}
break;
case LANCZOS:
interpolated = 0.0f;
norm = 0.0f;
new_x = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2];
new_x_int = new_x;
new_y = thread_data->distortion_coords[x * 2 * 3 + y * thread_data->width * 2 * 3 + rgb_index * 2 + 1];
new_y_int = new_y;
for (j = 0; j < 4; ++j)
for (i = 0; i < 4; ++i) {
if (new_x_int + i - 2 < 0 || new_x_int + i - 2 >= thread_data->width || new_y_int + j - 2 < 0 || new_y_int + j - 2 >= thread_data->height)
continue;
d = square(new_x - (new_x_int + i - 2)) * square(new_y - (new_y_int + j - 2));
if (d >= 4.0f)
continue;
d = thread_data->interpolation[(int)(d * LANCZOS_RESOLUTION)];
norm += d;
interpolated += thread_data->data_in[(new_x_int + i - 2) * 3 + rgb_index + (new_y_int + j - 2) * thread_data->linesize_in] * d;
}
if (norm == 0.0f) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
interpolated /= norm;
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = interpolated < 0.0f ? 0.0f : interpolated > 255.0f ? 255.0f : interpolated;
}
break;
}
} else if (thread_data->mode & GEOMETRY_DISTORTION) {
// geometry distortion correction was applied, correct distortion
switch(thread_data->interpolation_type) {
case NEAREST:
new_x_int = thread_data->distortion_coords[x * 2 + y * thread_data->width * 2] + 0.5f;
new_y_int = thread_data->distortion_coords[x * 2 + y * thread_data->width * 2 + 1] + 0.5f;
if (new_x_int < 0 || new_x_int >= thread_data->width || new_y_int < 0 || new_y_int >= thread_data->height) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = thread_data->data_in[new_x_int * 3 + rgb_index + new_y_int * thread_data->linesize_in];
}
break;
case LINEAR:
interpolated = 0.0f;
new_x = thread_data->distortion_coords[x * 2 + y * thread_data->width * 2];
new_x_int = new_x;
new_y = thread_data->distortion_coords[x * 2 + y * thread_data->width * 2 + 1];
new_y_int = new_y;
if (new_x_int < 0 || new_x_int + 1 >= thread_data->width || new_y_int < 0 || new_y_int + 1 >= thread_data->height) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] =
thread_data->data_in[ new_x_int * 3 + rgb_index + new_y_int * thread_data->linesize_in] * (new_x_int + 1 - new_x) * (new_y_int + 1 - new_y)
+ thread_data->data_in[(new_x_int + 1) * 3 + rgb_index + new_y_int * thread_data->linesize_in] * (new_x - new_x_int) * (new_y_int + 1 - new_y)
+ thread_data->data_in[ new_x_int * 3 + rgb_index + (new_y_int + 1) * thread_data->linesize_in] * (new_x_int + 1 - new_x) * (new_y - new_y_int)
+ thread_data->data_in[(new_x_int + 1) * 3 + rgb_index + (new_y_int + 1) * thread_data->linesize_in] * (new_x - new_x_int) * (new_y - new_y_int);
}
break;
case LANCZOS:
interpolated = 0.0f;
norm = 0.0f;
new_x = thread_data->distortion_coords[x * 2 + y * thread_data->width * 2];
new_x_int = new_x;
new_y = thread_data->distortion_coords[x * 2 + 1 + y * thread_data->width * 2];
new_y_int = new_y;
for (j = 0; j < 4; ++j)
for (i = 0; i < 4; ++i) {
if (new_x_int + i - 2 < 0 || new_x_int + i - 2 >= thread_data->width || new_y_int + j - 2 < 0 || new_y_int + j - 2 >= thread_data->height)
continue;
d = square(new_x - (new_x_int + i - 2)) * square(new_y - (new_y_int + j - 2));
if (d >= 4.0f)
continue;
d = thread_data->interpolation[(int)(d * LANCZOS_RESOLUTION)];
norm += d;
interpolated += thread_data->data_in[(new_x_int + i - 2) * 3 + rgb_index + (new_y_int + j - 2) * thread_data->linesize_in] * d;
}
if (norm == 0.0f) {
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = 0;
} else {
interpolated /= norm;
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = interpolated < 0.0f ? 0.0f : interpolated > 255.0f ? 255.0f : interpolated;
}
break;
}
} else {
// no distortion correction was applied
thread_data->data_out[x * 3 + rgb_index + y * thread_data->linesize_out] = thread_data->data_in[x * 3 + rgb_index + y * thread_data->linesize_in];
}
}
return 0;
}
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
AVFilterContext *ctx = inlink->dst;
LensfunContext *lensfun = ctx->priv;
AVFilterLink *outlink = ctx->outputs[0];
AVFrame *out;
VignettingThreadData vignetting_thread_data;
DistortionCorrectionThreadData distortion_correction_thread_data;
if (lensfun->mode & VIGNETTING) {
av_frame_make_writable(in);
vignetting_thread_data = (VignettingThreadData) {
.width = inlink->w,
.height = inlink->h,
.data_in = in->data[0],
.linesize_in = in->linesize[0],
.pixel_composition = LF_CR_3(RED, GREEN, BLUE),
.modifier = lensfun->modifier
};
ctx->internal->execute(ctx,
vignetting_filter_slice,
&vignetting_thread_data,
NULL,
FFMIN(outlink->h, ctx->graph->nb_threads));
}
if (lensfun->mode & (GEOMETRY_DISTORTION | SUBPIXEL_DISTORTION)) {
out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
if (!out) {
av_frame_free(&in);
return AVERROR(ENOMEM);
}
av_frame_copy_props(out, in);
distortion_correction_thread_data = (DistortionCorrectionThreadData) {
.width = inlink->w,
.height = inlink->h,
.distortion_coords = lensfun->distortion_coords,
.data_in = in->data[0],
.data_out = out->data[0],
.linesize_in = in->linesize[0],
.linesize_out = out->linesize[0],
.interpolation = lensfun->interpolation,
.mode = lensfun->mode,
.interpolation_type = lensfun->interpolation_type
};
ctx->internal->execute(ctx,
distortion_correction_filter_slice,
&distortion_correction_thread_data,
NULL,
FFMIN(outlink->h, ctx->graph->nb_threads));
av_frame_free(&in);
return ff_filter_frame(outlink, out);
} else {
return ff_filter_frame(outlink, in);
}
}
static av_cold void uninit(AVFilterContext *ctx)
{
LensfunContext *lensfun = ctx->priv;
if (lensfun->camera)
lf_camera_destroy(lensfun->camera);
if (lensfun->lens)
lf_lens_destroy(lensfun->lens);
if (lensfun->modifier)
lf_modifier_destroy(lensfun->modifier);
if (lensfun->distortion_coords)
av_free(lensfun->distortion_coords);
if (lensfun->interpolation)
av_free(lensfun->interpolation);
}
static const AVFilterPad lensfun_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_props,
.filter_frame = filter_frame,
},
{ NULL }
};
static const AVFilterPad lensfun_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
},
{ NULL }
};
AVFilter ff_vf_lensfun = {
.name = "lensfun",
.description = NULL_IF_CONFIG_SMALL("Apply correction to an image based on info derived from the lensfun database."),
.priv_size = sizeof(LensfunContext),
.init = init,
.uninit = uninit,
.query_formats = query_formats,
.inputs = lensfun_inputs,
.outputs = lensfun_outputs,
.priv_class = &lensfun_class,
.flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS,
};