1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-02 03:06:28 +02:00
FFmpeg/libavfilter/vf_showinfo.c
Limin Wang 6866c54ee4 avfilter/showinfo: support Content Light Level information
show real information instead of the unknown side data type message for HDR10 stream

Signed-off-by: Limin Wang <lance.lmwang@gmail.com>
Signed-off-by: James Almer <jamrial@gmail.com>
2019-08-17 14:47:13 -03:00

380 lines
13 KiB
C

/*
* Copyright (c) 2011 Stefano Sabatini
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file
* filter for showing textual video frame information
*/
#include <inttypes.h>
#include "libavutil/adler32.h"
#include "libavutil/display.h"
#include "libavutil/imgutils.h"
#include "libavutil/internal.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "libavutil/spherical.h"
#include "libavutil/stereo3d.h"
#include "libavutil/timestamp.h"
#include "libavutil/timecode.h"
#include "libavutil/mastering_display_metadata.h"
#include "avfilter.h"
#include "internal.h"
#include "video.h"
typedef struct ShowInfoContext {
const AVClass *class;
int calculate_checksums;
} ShowInfoContext;
#define OFFSET(x) offsetof(ShowInfoContext, x)
#define VF AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
static const AVOption showinfo_options[] = {
{ "checksum", "calculate checksums", OFFSET(calculate_checksums), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, VF },
{ NULL }
};
AVFILTER_DEFINE_CLASS(showinfo);
static void dump_spherical(AVFilterContext *ctx, AVFrame *frame, AVFrameSideData *sd)
{
AVSphericalMapping *spherical = (AVSphericalMapping *)sd->data;
double yaw, pitch, roll;
av_log(ctx, AV_LOG_INFO, "spherical information: ");
if (sd->size < sizeof(*spherical)) {
av_log(ctx, AV_LOG_ERROR, "invalid data");
return;
}
if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR)
av_log(ctx, AV_LOG_INFO, "equirectangular ");
else if (spherical->projection == AV_SPHERICAL_CUBEMAP)
av_log(ctx, AV_LOG_INFO, "cubemap ");
else if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR_TILE)
av_log(ctx, AV_LOG_INFO, "tiled equirectangular ");
else {
av_log(ctx, AV_LOG_WARNING, "unknown");
return;
}
yaw = ((double)spherical->yaw) / (1 << 16);
pitch = ((double)spherical->pitch) / (1 << 16);
roll = ((double)spherical->roll) / (1 << 16);
av_log(ctx, AV_LOG_INFO, "(%f/%f/%f) ", yaw, pitch, roll);
if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR_TILE) {
size_t l, t, r, b;
av_spherical_tile_bounds(spherical, frame->width, frame->height,
&l, &t, &r, &b);
av_log(ctx, AV_LOG_INFO,
"[%"SIZE_SPECIFIER", %"SIZE_SPECIFIER", %"SIZE_SPECIFIER", %"SIZE_SPECIFIER"] ",
l, t, r, b);
} else if (spherical->projection == AV_SPHERICAL_CUBEMAP) {
av_log(ctx, AV_LOG_INFO, "[pad %"PRIu32"] ", spherical->padding);
}
}
static void dump_stereo3d(AVFilterContext *ctx, AVFrameSideData *sd)
{
AVStereo3D *stereo;
av_log(ctx, AV_LOG_INFO, "stereoscopic information: ");
if (sd->size < sizeof(*stereo)) {
av_log(ctx, AV_LOG_ERROR, "invalid data");
return;
}
stereo = (AVStereo3D *)sd->data;
av_log(ctx, AV_LOG_INFO, "type - %s", av_stereo3d_type_name(stereo->type));
if (stereo->flags & AV_STEREO3D_FLAG_INVERT)
av_log(ctx, AV_LOG_INFO, " (inverted)");
}
static void dump_roi(AVFilterContext *ctx, AVFrameSideData *sd)
{
int nb_rois;
const AVRegionOfInterest *roi;
uint32_t roi_size;
roi = (const AVRegionOfInterest *)sd->data;
roi_size = roi->self_size;
if (!roi_size || sd->size % roi_size != 0) {
av_log(ctx, AV_LOG_ERROR, "Invalid AVRegionOfInterest.self_size.");
return;
}
nb_rois = sd->size / roi_size;
av_log(ctx, AV_LOG_INFO, "Regions Of Interest(RoI) information: ");
for (int i = 0; i < nb_rois; i++) {
roi = (const AVRegionOfInterest *)(sd->data + roi_size * i);
av_log(ctx, AV_LOG_INFO, "index: %d, region: (%d, %d)/(%d, %d), qp offset: %d/%d.\n",
i, roi->left, roi->top, roi->right, roi->bottom, roi->qoffset.num, roi->qoffset.den);
}
}
static void dump_mastering_display(AVFilterContext *ctx, AVFrameSideData *sd)
{
AVMasteringDisplayMetadata *mastering_display;
av_log(ctx, AV_LOG_INFO, "mastering display: ");
if (sd->size < sizeof(*mastering_display)) {
av_log(ctx, AV_LOG_ERROR, "invalid data");
return;
}
mastering_display = (AVMasteringDisplayMetadata *)sd->data;
av_log(ctx, AV_LOG_INFO, "has_primaries:%d has_luminance:%d "
"r(%5.4f,%5.4f) g(%5.4f,%5.4f) b(%5.4f %5.4f) wp(%5.4f, %5.4f) "
"min_luminance=%f, max_luminance=%f",
mastering_display->has_primaries, mastering_display->has_luminance,
av_q2d(mastering_display->display_primaries[0][0]),
av_q2d(mastering_display->display_primaries[0][1]),
av_q2d(mastering_display->display_primaries[1][0]),
av_q2d(mastering_display->display_primaries[1][1]),
av_q2d(mastering_display->display_primaries[2][0]),
av_q2d(mastering_display->display_primaries[2][1]),
av_q2d(mastering_display->white_point[0]), av_q2d(mastering_display->white_point[1]),
av_q2d(mastering_display->min_luminance), av_q2d(mastering_display->max_luminance));
}
static void dump_content_light_metadata(AVFilterContext *ctx, AVFrameSideData *sd)
{
AVContentLightMetadata* metadata = (AVContentLightMetadata*)sd->data;
av_log(ctx, AV_LOG_INFO, "Content Light Level information: "
"MaxCLL=%d, MaxFALL=%d",
metadata->MaxCLL, metadata->MaxFALL);
}
static void dump_color_property(AVFilterContext *ctx, AVFrame *frame)
{
const char *color_range_str = av_color_range_name(frame->color_range);
const char *colorspace_str = av_color_space_name(frame->colorspace);
const char *color_primaries_str = av_color_primaries_name(frame->color_primaries);
const char *color_trc_str = av_color_transfer_name(frame->color_trc);
if (!color_range_str || frame->color_range == AVCOL_RANGE_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, "color_range:unknown");
} else {
av_log(ctx, AV_LOG_INFO, "color_range:%s", color_range_str);
}
if (!colorspace_str || frame->colorspace == AVCOL_SPC_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_space:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_space:%s", colorspace_str);
}
if (!color_primaries_str || frame->color_primaries == AVCOL_PRI_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_primaries:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_primaries:%s", color_primaries_str);
}
if (!color_trc_str || frame->color_trc == AVCOL_TRC_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_trc:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_trc:%s", color_trc_str);
}
av_log(ctx, AV_LOG_INFO, "\n");
}
static void update_sample_stats(const uint8_t *src, int len, int64_t *sum, int64_t *sum2)
{
int i;
for (i = 0; i < len; i++) {
*sum += src[i];
*sum2 += src[i] * src[i];
}
}
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
AVFilterContext *ctx = inlink->dst;
ShowInfoContext *s = ctx->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
uint32_t plane_checksum[4] = {0}, checksum = 0;
int64_t sum[4] = {0}, sum2[4] = {0};
int32_t pixelcount[4] = {0};
int i, plane, vsub = desc->log2_chroma_h;
for (plane = 0; plane < 4 && s->calculate_checksums && frame->data[plane] && frame->linesize[plane]; plane++) {
uint8_t *data = frame->data[plane];
int h = plane == 1 || plane == 2 ? AV_CEIL_RSHIFT(inlink->h, vsub) : inlink->h;
int linesize = av_image_get_linesize(frame->format, frame->width, plane);
if (linesize < 0)
return linesize;
for (i = 0; i < h; i++) {
plane_checksum[plane] = av_adler32_update(plane_checksum[plane], data, linesize);
checksum = av_adler32_update(checksum, data, linesize);
update_sample_stats(data, linesize, sum+plane, sum2+plane);
pixelcount[plane] += linesize;
data += frame->linesize[plane];
}
}
av_log(ctx, AV_LOG_INFO,
"n:%4"PRId64" pts:%7s pts_time:%-7s pos:%9"PRId64" "
"fmt:%s sar:%d/%d s:%dx%d i:%c iskey:%d type:%c ",
inlink->frame_count_out,
av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base), frame->pkt_pos,
desc->name,
frame->sample_aspect_ratio.num, frame->sample_aspect_ratio.den,
frame->width, frame->height,
!frame->interlaced_frame ? 'P' : /* Progressive */
frame->top_field_first ? 'T' : 'B', /* Top / Bottom */
frame->key_frame,
av_get_picture_type_char(frame->pict_type));
if (s->calculate_checksums) {
av_log(ctx, AV_LOG_INFO,
"checksum:%08"PRIX32" plane_checksum:[%08"PRIX32,
checksum, plane_checksum[0]);
for (plane = 1; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, " %08"PRIX32, plane_checksum[plane]);
av_log(ctx, AV_LOG_INFO, "] mean:[");
for (plane = 0; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, "%"PRId64" ", (sum[plane] + pixelcount[plane]/2) / pixelcount[plane]);
av_log(ctx, AV_LOG_INFO, "\b] stdev:[");
for (plane = 0; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, "%3.1f ",
sqrt((sum2[plane] - sum[plane]*(double)sum[plane]/pixelcount[plane])/pixelcount[plane]));
av_log(ctx, AV_LOG_INFO, "\b]");
}
av_log(ctx, AV_LOG_INFO, "\n");
for (i = 0; i < frame->nb_side_data; i++) {
AVFrameSideData *sd = frame->side_data[i];
av_log(ctx, AV_LOG_INFO, " side data - ");
switch (sd->type) {
case AV_FRAME_DATA_PANSCAN:
av_log(ctx, AV_LOG_INFO, "pan/scan");
break;
case AV_FRAME_DATA_A53_CC:
av_log(ctx, AV_LOG_INFO, "A/53 closed captions (%d bytes)", sd->size);
break;
case AV_FRAME_DATA_SPHERICAL:
dump_spherical(ctx, frame, sd);
break;
case AV_FRAME_DATA_STEREO3D:
dump_stereo3d(ctx, sd);
break;
case AV_FRAME_DATA_S12M_TIMECODE: {
uint32_t *tc = (uint32_t*)sd->data;
for (int j = 1; j <= tc[0]; j++) {
char tcbuf[AV_TIMECODE_STR_SIZE];
av_timecode_make_smpte_tc_string(tcbuf, tc[j], 0);
av_log(ctx, AV_LOG_INFO, "timecode - %s%s", tcbuf, j != tc[0] ? ", " : "");
}
break;
}
case AV_FRAME_DATA_DISPLAYMATRIX:
av_log(ctx, AV_LOG_INFO, "displaymatrix: rotation of %.2f degrees",
av_display_rotation_get((int32_t *)sd->data));
break;
case AV_FRAME_DATA_AFD:
av_log(ctx, AV_LOG_INFO, "afd: value of %"PRIu8, sd->data[0]);
break;
case AV_FRAME_DATA_REGIONS_OF_INTEREST:
dump_roi(ctx, sd);
break;
case AV_FRAME_DATA_MASTERING_DISPLAY_METADATA:
dump_mastering_display(ctx, sd);
break;
case AV_FRAME_DATA_CONTENT_LIGHT_LEVEL:
dump_content_light_metadata(ctx, sd);
break;
default:
av_log(ctx, AV_LOG_WARNING, "unknown side data type %d (%d bytes)",
sd->type, sd->size);
break;
}
av_log(ctx, AV_LOG_INFO, "\n");
}
dump_color_property(ctx, frame);
return ff_filter_frame(inlink->dst->outputs[0], frame);
}
static int config_props(AVFilterContext *ctx, AVFilterLink *link, int is_out)
{
av_log(ctx, AV_LOG_INFO, "config %s time_base: %d/%d, frame_rate: %d/%d\n",
is_out ? "out" : "in",
link->time_base.num, link->time_base.den,
link->frame_rate.num, link->frame_rate.den);
return 0;
}
static int config_props_in(AVFilterLink *link)
{
AVFilterContext *ctx = link->dst;
return config_props(ctx, link, 0);
}
static int config_props_out(AVFilterLink *link)
{
AVFilterContext *ctx = link->src;
return config_props(ctx, link, 1);
}
static const AVFilterPad avfilter_vf_showinfo_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = filter_frame,
.config_props = config_props_in,
},
{ NULL }
};
static const AVFilterPad avfilter_vf_showinfo_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_props_out,
},
{ NULL }
};
AVFilter ff_vf_showinfo = {
.name = "showinfo",
.description = NULL_IF_CONFIG_SMALL("Show textual information for each video frame."),
.inputs = avfilter_vf_showinfo_inputs,
.outputs = avfilter_vf_showinfo_outputs,
.priv_size = sizeof(ShowInfoContext),
.priv_class = &showinfo_class,
};