mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-28 20:53:54 +02:00
343 lines
12 KiB
C
343 lines
12 KiB
C
/*
|
|
* Copyright (c) 2017 Ronald S. Bultje <rsbultje@gmail.com>
|
|
* Copyright (c) 2017 Ashish Pratap Singh <ashk43712@gmail.com>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
* Calculate the VMAF between two input videos.
|
|
*/
|
|
|
|
#include <pthread.h>
|
|
#include <libvmaf.h>
|
|
#include "libavutil/avstring.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/pixdesc.h"
|
|
#include "avfilter.h"
|
|
#include "drawutils.h"
|
|
#include "formats.h"
|
|
#include "framesync.h"
|
|
#include "internal.h"
|
|
#include "video.h"
|
|
|
|
typedef struct LIBVMAFContext {
|
|
const AVClass *class;
|
|
FFFrameSync fs;
|
|
const AVPixFmtDescriptor *desc;
|
|
char *format;
|
|
int width;
|
|
int height;
|
|
double vmaf_score;
|
|
pthread_t vmaf_thread;
|
|
pthread_mutex_t lock;
|
|
pthread_cond_t cond;
|
|
int eof;
|
|
AVFrame *gmain;
|
|
AVFrame *gref;
|
|
int frame_set;
|
|
char *model_path;
|
|
char *log_path;
|
|
char *log_fmt;
|
|
int disable_clip;
|
|
int disable_avx;
|
|
int enable_transform;
|
|
int phone_model;
|
|
int psnr;
|
|
int ssim;
|
|
int ms_ssim;
|
|
char *pool;
|
|
} LIBVMAFContext;
|
|
|
|
#define OFFSET(x) offsetof(LIBVMAFContext, x)
|
|
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
|
|
|
|
static const AVOption libvmaf_options[] = {
|
|
{"model_path", "Set the model to be used for computing vmaf.", OFFSET(model_path), AV_OPT_TYPE_STRING, {.str="/usr/local/share/model/vmaf_v0.6.1.pkl"}, 0, 1, FLAGS},
|
|
{"log_path", "Set the file path to be used to store logs.", OFFSET(log_path), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 1, FLAGS},
|
|
{"log_fmt", "Set the format of the log (xml or json).", OFFSET(log_fmt), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 1, FLAGS},
|
|
{"enable_transform", "Enables transform for computing vmaf.", OFFSET(enable_transform), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS},
|
|
{"phone_model", "Invokes the phone model that will generate higher VMAF scores.", OFFSET(phone_model), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS},
|
|
{"psnr", "Enables computing psnr along with vmaf.", OFFSET(psnr), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS},
|
|
{"ssim", "Enables computing ssim along with vmaf.", OFFSET(ssim), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS},
|
|
{"ms_ssim", "Enables computing ms-ssim along with vmaf.", OFFSET(ms_ssim), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS},
|
|
{"pool", "Set the pool method to be used for computing vmaf.", OFFSET(pool), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 1, FLAGS},
|
|
{ NULL }
|
|
};
|
|
|
|
FRAMESYNC_DEFINE_CLASS(libvmaf, LIBVMAFContext, fs);
|
|
|
|
#define read_frame_fn(type, bits) \
|
|
static int read_frame_##bits##bit(float *ref_data, float *main_data, \
|
|
float *temp_data, int stride, void *ctx) \
|
|
{ \
|
|
LIBVMAFContext *s = (LIBVMAFContext *) ctx; \
|
|
int ret; \
|
|
\
|
|
pthread_mutex_lock(&s->lock); \
|
|
\
|
|
while (!s->frame_set && !s->eof) { \
|
|
pthread_cond_wait(&s->cond, &s->lock); \
|
|
} \
|
|
\
|
|
if (s->frame_set) { \
|
|
int ref_stride = s->gref->linesize[0]; \
|
|
int main_stride = s->gmain->linesize[0]; \
|
|
\
|
|
const type *ref_ptr = (const type *) s->gref->data[0]; \
|
|
const type *main_ptr = (const type *) s->gmain->data[0]; \
|
|
\
|
|
float *ptr = ref_data; \
|
|
\
|
|
int h = s->height; \
|
|
int w = s->width; \
|
|
\
|
|
int i,j; \
|
|
\
|
|
for (i = 0; i < h; i++) { \
|
|
for ( j = 0; j < w; j++) { \
|
|
ptr[j] = (float)ref_ptr[j]; \
|
|
} \
|
|
ref_ptr += ref_stride / sizeof(*ref_ptr); \
|
|
ptr += stride / sizeof(*ptr); \
|
|
} \
|
|
\
|
|
ptr = main_data; \
|
|
\
|
|
for (i = 0; i < h; i++) { \
|
|
for (j = 0; j < w; j++) { \
|
|
ptr[j] = (float)main_ptr[j]; \
|
|
} \
|
|
main_ptr += main_stride / sizeof(*main_ptr); \
|
|
ptr += stride / sizeof(*ptr); \
|
|
} \
|
|
} \
|
|
\
|
|
ret = !s->frame_set; \
|
|
\
|
|
s->frame_set = 0; \
|
|
\
|
|
pthread_cond_signal(&s->cond); \
|
|
pthread_mutex_unlock(&s->lock); \
|
|
\
|
|
if (ret) { \
|
|
return 2; \
|
|
} \
|
|
\
|
|
return 0; \
|
|
}
|
|
|
|
read_frame_fn(uint8_t, 8);
|
|
read_frame_fn(uint16_t, 10);
|
|
|
|
static void compute_vmaf_score(LIBVMAFContext *s)
|
|
{
|
|
int (*read_frame)(float *ref_data, float *main_data, float *temp_data,
|
|
int stride, void *ctx);
|
|
|
|
if (s->desc->comp[0].depth <= 8) {
|
|
read_frame = read_frame_8bit;
|
|
} else {
|
|
read_frame = read_frame_10bit;
|
|
}
|
|
|
|
s->vmaf_score = compute_vmaf(s->format, s->width, s->height, read_frame, s,
|
|
s->model_path, s->log_path, s->log_fmt, 0, 0,
|
|
s->enable_transform, s->phone_model, s->psnr,
|
|
s->ssim, s->ms_ssim, s->pool);
|
|
}
|
|
|
|
static void *call_vmaf(void *ctx)
|
|
{
|
|
LIBVMAFContext *s = (LIBVMAFContext *) ctx;
|
|
compute_vmaf_score(s);
|
|
av_log(ctx, AV_LOG_INFO, "VMAF score: %f\n",s->vmaf_score);
|
|
pthread_exit(NULL);
|
|
}
|
|
|
|
static int do_vmaf(FFFrameSync *fs)
|
|
{
|
|
AVFilterContext *ctx = fs->parent;
|
|
LIBVMAFContext *s = ctx->priv;
|
|
AVFrame *main, *ref;
|
|
int ret;
|
|
|
|
ret = ff_framesync_dualinput_get(fs, &main, &ref);
|
|
if (ret < 0)
|
|
return ret;
|
|
if (!ref)
|
|
return ff_filter_frame(ctx->outputs[0], main);
|
|
|
|
pthread_mutex_lock(&s->lock);
|
|
|
|
while (s->frame_set != 0) {
|
|
pthread_cond_wait(&s->cond, &s->lock);
|
|
}
|
|
|
|
av_frame_ref(s->gref, ref);
|
|
av_frame_ref(s->gmain, main);
|
|
|
|
s->frame_set = 1;
|
|
|
|
pthread_cond_signal(&s->cond);
|
|
pthread_mutex_unlock(&s->lock);
|
|
|
|
return ff_filter_frame(ctx->outputs[0], main);
|
|
}
|
|
|
|
static av_cold int init(AVFilterContext *ctx)
|
|
{
|
|
LIBVMAFContext *s = ctx->priv;
|
|
|
|
s->gref = av_frame_alloc();
|
|
s->gmain = av_frame_alloc();
|
|
|
|
pthread_mutex_init(&s->lock, NULL);
|
|
pthread_cond_init (&s->cond, NULL);
|
|
|
|
s->fs.on_event = do_vmaf;
|
|
return 0;
|
|
}
|
|
|
|
static int query_formats(AVFilterContext *ctx)
|
|
{
|
|
static const enum AVPixelFormat pix_fmts[] = {
|
|
AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P,
|
|
AV_PIX_FMT_YUV444P10LE, AV_PIX_FMT_YUV422P10LE, AV_PIX_FMT_YUV420P10LE,
|
|
AV_PIX_FMT_NONE
|
|
};
|
|
|
|
AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts);
|
|
if (!fmts_list)
|
|
return AVERROR(ENOMEM);
|
|
return ff_set_common_formats(ctx, fmts_list);
|
|
}
|
|
|
|
|
|
static int config_input_ref(AVFilterLink *inlink)
|
|
{
|
|
AVFilterContext *ctx = inlink->dst;
|
|
LIBVMAFContext *s = ctx->priv;
|
|
int th;
|
|
|
|
if (ctx->inputs[0]->w != ctx->inputs[1]->w ||
|
|
ctx->inputs[0]->h != ctx->inputs[1]->h) {
|
|
av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
if (ctx->inputs[0]->format != ctx->inputs[1]->format) {
|
|
av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
s->desc = av_pix_fmt_desc_get(inlink->format);
|
|
s->width = ctx->inputs[0]->w;
|
|
s->height = ctx->inputs[0]->h;
|
|
|
|
th = pthread_create(&s->vmaf_thread, NULL, call_vmaf, (void *) s);
|
|
if (th) {
|
|
av_log(ctx, AV_LOG_ERROR, "Thread creation failed.\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int config_output(AVFilterLink *outlink)
|
|
{
|
|
AVFilterContext *ctx = outlink->src;
|
|
LIBVMAFContext *s = ctx->priv;
|
|
AVFilterLink *mainlink = ctx->inputs[0];
|
|
int ret;
|
|
|
|
ret = ff_framesync_init_dualinput(&s->fs, ctx);
|
|
if (ret < 0)
|
|
return ret;
|
|
outlink->w = mainlink->w;
|
|
outlink->h = mainlink->h;
|
|
outlink->time_base = mainlink->time_base;
|
|
outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio;
|
|
outlink->frame_rate = mainlink->frame_rate;
|
|
if ((ret = ff_framesync_configure(&s->fs)) < 0)
|
|
return ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int activate(AVFilterContext *ctx)
|
|
{
|
|
LIBVMAFContext *s = ctx->priv;
|
|
return ff_framesync_activate(&s->fs);
|
|
}
|
|
|
|
static av_cold void uninit(AVFilterContext *ctx)
|
|
{
|
|
LIBVMAFContext *s = ctx->priv;
|
|
|
|
ff_framesync_uninit(&s->fs);
|
|
|
|
pthread_mutex_lock(&s->lock);
|
|
s->eof = 1;
|
|
pthread_cond_signal(&s->cond);
|
|
pthread_mutex_unlock(&s->lock);
|
|
|
|
pthread_join(s->vmaf_thread, NULL);
|
|
|
|
av_frame_free(&s->gref);
|
|
av_frame_free(&s->gmain);
|
|
|
|
pthread_mutex_destroy(&s->lock);
|
|
pthread_cond_destroy(&s->cond);
|
|
}
|
|
|
|
static const AVFilterPad libvmaf_inputs[] = {
|
|
{
|
|
.name = "main",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
},{
|
|
.name = "reference",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.config_props = config_input_ref,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVFilterPad libvmaf_outputs[] = {
|
|
{
|
|
.name = "default",
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.config_props = config_output,
|
|
},
|
|
{ NULL }
|
|
};
|
|
|
|
AVFilter ff_vf_libvmaf = {
|
|
.name = "libvmaf",
|
|
.description = NULL_IF_CONFIG_SMALL("Calculate the VMAF between two video streams."),
|
|
.preinit = libvmaf_framesync_preinit,
|
|
.init = init,
|
|
.uninit = uninit,
|
|
.query_formats = query_formats,
|
|
.activate = activate,
|
|
.priv_size = sizeof(LIBVMAFContext),
|
|
.priv_class = &libvmaf_class,
|
|
.inputs = libvmaf_inputs,
|
|
.outputs = libvmaf_outputs,
|
|
};
|