1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-23 12:43:46 +02:00
FFmpeg/libavfilter/vf_showinfo.c
Andreas Rheinhardt 8be701d9f7 avfilter/avfilter: Add numbers of (in|out)pads directly to AVFilter
Up until now, an AVFilter's lists of input and output AVFilterPads
were terminated by a sentinel and the only way to get the length
of these lists was by using avfilter_pad_count(). This has two
drawbacks: first, sizeof(AVFilterPad) is not negligible
(i.e. 64B on 64bit systems); second, getting the size involves
a function call instead of just reading the data.

This commit therefore changes this. The sentinels are removed and new
private fields nb_inputs and nb_outputs are added to AVFilter that
contain the number of elements of the respective AVFilterPad array.

Given that AVFilter.(in|out)puts are the only arrays of zero-terminated
AVFilterPads an API user has access to (AVFilterContext.(in|out)put_pads
are not zero-terminated and they already have a size field) the argument
to avfilter_pad_count() is always one of these lists, so it just has to
find the filter the list belongs to and read said number. This is slower
than before, but a replacement function that just reads the internal numbers
that users are expected to switch to will be added soon; and furthermore,
avfilter_pad_count() is probably never called in hot loops anyway.

This saves about 49KiB from the binary; notice that these sentinels are
not in .bss despite being zeroed: they are in .data.rel.ro due to the
non-sentinels.

Reviewed-by: Nicolas George <george@nsup.org>
Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
2021-08-20 12:53:58 +02:00

682 lines
27 KiB
C

/*
* Copyright (c) 2011 Stefano Sabatini
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
/**
* @file
* filter for showing textual video frame information
*/
#include <inttypes.h>
#include "libavutil/bswap.h"
#include "libavutil/adler32.h"
#include "libavutil/display.h"
#include "libavutil/imgutils.h"
#include "libavutil/internal.h"
#include "libavutil/film_grain_params.h"
#include "libavutil/hdr_dynamic_metadata.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "libavutil/spherical.h"
#include "libavutil/stereo3d.h"
#include "libavutil/timestamp.h"
#include "libavutil/timecode.h"
#include "libavutil/mastering_display_metadata.h"
#include "libavutil/video_enc_params.h"
#include "libavutil/detection_bbox.h"
#include "avfilter.h"
#include "internal.h"
#include "video.h"
typedef struct ShowInfoContext {
const AVClass *class;
int calculate_checksums;
} ShowInfoContext;
#define OFFSET(x) offsetof(ShowInfoContext, x)
#define VF AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
static const AVOption showinfo_options[] = {
{ "checksum", "calculate checksums", OFFSET(calculate_checksums), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, VF },
{ NULL }
};
AVFILTER_DEFINE_CLASS(showinfo);
static void dump_spherical(AVFilterContext *ctx, AVFrame *frame, const AVFrameSideData *sd)
{
const AVSphericalMapping *spherical = (const AVSphericalMapping *)sd->data;
double yaw, pitch, roll;
av_log(ctx, AV_LOG_INFO, "spherical information: ");
if (sd->size < sizeof(*spherical)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR)
av_log(ctx, AV_LOG_INFO, "equirectangular ");
else if (spherical->projection == AV_SPHERICAL_CUBEMAP)
av_log(ctx, AV_LOG_INFO, "cubemap ");
else if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR_TILE)
av_log(ctx, AV_LOG_INFO, "tiled equirectangular ");
else {
av_log(ctx, AV_LOG_WARNING, "unknown\n");
return;
}
yaw = ((double)spherical->yaw) / (1 << 16);
pitch = ((double)spherical->pitch) / (1 << 16);
roll = ((double)spherical->roll) / (1 << 16);
av_log(ctx, AV_LOG_INFO, "(%f/%f/%f) ", yaw, pitch, roll);
if (spherical->projection == AV_SPHERICAL_EQUIRECTANGULAR_TILE) {
size_t l, t, r, b;
av_spherical_tile_bounds(spherical, frame->width, frame->height,
&l, &t, &r, &b);
av_log(ctx, AV_LOG_INFO,
"[%"SIZE_SPECIFIER", %"SIZE_SPECIFIER", %"SIZE_SPECIFIER", %"SIZE_SPECIFIER"] ",
l, t, r, b);
} else if (spherical->projection == AV_SPHERICAL_CUBEMAP) {
av_log(ctx, AV_LOG_INFO, "[pad %"PRIu32"] ", spherical->padding);
}
}
static void dump_stereo3d(AVFilterContext *ctx, const AVFrameSideData *sd)
{
const AVStereo3D *stereo;
av_log(ctx, AV_LOG_INFO, "stereoscopic information: ");
if (sd->size < sizeof(*stereo)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
stereo = (const AVStereo3D *)sd->data;
av_log(ctx, AV_LOG_INFO, "type - %s", av_stereo3d_type_name(stereo->type));
if (stereo->flags & AV_STEREO3D_FLAG_INVERT)
av_log(ctx, AV_LOG_INFO, " (inverted)");
}
static void dump_s12m_timecode(AVFilterContext *ctx, AVRational frame_rate, const AVFrameSideData *sd)
{
const uint32_t *tc = (const uint32_t *)sd->data;
if ((sd->size != sizeof(uint32_t) * 4) || (tc[0] > 3)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
for (int j = 1; j <= tc[0]; j++) {
char tcbuf[AV_TIMECODE_STR_SIZE];
av_timecode_make_smpte_tc_string2(tcbuf, frame_rate, tc[j], 0, 0);
av_log(ctx, AV_LOG_INFO, "timecode - %s%s", tcbuf, j != tc[0] ? ", " : "");
}
}
static void dump_roi(AVFilterContext *ctx, const AVFrameSideData *sd)
{
int nb_rois;
const AVRegionOfInterest *roi;
uint32_t roi_size;
roi = (const AVRegionOfInterest *)sd->data;
roi_size = roi->self_size;
if (!roi_size || sd->size % roi_size != 0) {
av_log(ctx, AV_LOG_ERROR, "Invalid AVRegionOfInterest.self_size.\n");
return;
}
nb_rois = sd->size / roi_size;
av_log(ctx, AV_LOG_INFO, "Regions Of Interest(RoI) information: ");
for (int i = 0; i < nb_rois; i++) {
roi = (const AVRegionOfInterest *)(sd->data + roi_size * i);
av_log(ctx, AV_LOG_INFO, "index: %d, region: (%d, %d)/(%d, %d), qp offset: %d/%d.\n",
i, roi->left, roi->top, roi->right, roi->bottom, roi->qoffset.num, roi->qoffset.den);
}
}
static void dump_detection_bbox(AVFilterContext *ctx, const AVFrameSideData *sd)
{
int nb_bboxes;
const AVDetectionBBoxHeader *header;
const AVDetectionBBox *bbox;
header = (const AVDetectionBBoxHeader *)sd->data;
nb_bboxes = header->nb_bboxes;
av_log(ctx, AV_LOG_INFO, "detection bounding boxes:\n");
av_log(ctx, AV_LOG_INFO, "source: %s\n", header->source);
for (int i = 0; i < nb_bboxes; i++) {
bbox = av_get_detection_bbox(header, i);
av_log(ctx, AV_LOG_INFO, "index: %d,\tregion: (%d, %d) -> (%d, %d), label: %s, confidence: %d/%d.\n",
i, bbox->x, bbox->y, bbox->x + bbox->w, bbox->y + bbox->h,
bbox->detect_label, bbox->detect_confidence.num, bbox->detect_confidence.den);
if (bbox->classify_count > 0) {
for (int j = 0; j < bbox->classify_count; j++) {
av_log(ctx, AV_LOG_INFO, "\t\tclassify: label: %s, confidence: %d/%d.\n",
bbox->classify_labels[j], bbox->classify_confidences[j].num, bbox->classify_confidences[j].den);
}
}
}
}
static void dump_mastering_display(AVFilterContext *ctx, const AVFrameSideData *sd)
{
const AVMasteringDisplayMetadata *mastering_display;
av_log(ctx, AV_LOG_INFO, "mastering display: ");
if (sd->size < sizeof(*mastering_display)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
mastering_display = (const AVMasteringDisplayMetadata *)sd->data;
av_log(ctx, AV_LOG_INFO, "has_primaries:%d has_luminance:%d "
"r(%5.4f,%5.4f) g(%5.4f,%5.4f) b(%5.4f %5.4f) wp(%5.4f, %5.4f) "
"min_luminance=%f, max_luminance=%f",
mastering_display->has_primaries, mastering_display->has_luminance,
av_q2d(mastering_display->display_primaries[0][0]),
av_q2d(mastering_display->display_primaries[0][1]),
av_q2d(mastering_display->display_primaries[1][0]),
av_q2d(mastering_display->display_primaries[1][1]),
av_q2d(mastering_display->display_primaries[2][0]),
av_q2d(mastering_display->display_primaries[2][1]),
av_q2d(mastering_display->white_point[0]), av_q2d(mastering_display->white_point[1]),
av_q2d(mastering_display->min_luminance), av_q2d(mastering_display->max_luminance));
}
static void dump_dynamic_hdr_plus(AVFilterContext *ctx, AVFrameSideData *sd)
{
AVDynamicHDRPlus *hdr_plus;
av_log(ctx, AV_LOG_INFO, "HDR10+ metadata: ");
if (sd->size < sizeof(*hdr_plus)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
hdr_plus = (AVDynamicHDRPlus *)sd->data;
av_log(ctx, AV_LOG_INFO, "application version: %d, ", hdr_plus->application_version);
av_log(ctx, AV_LOG_INFO, "num_windows: %d, ", hdr_plus->num_windows);
for (int w = 1; w < hdr_plus->num_windows; w++) {
AVHDRPlusColorTransformParams *params = &hdr_plus->params[w];
av_log(ctx, AV_LOG_INFO, "window %d { ", w);
av_log(ctx, AV_LOG_INFO, "window_upper_left_corner: (%5.4f,%5.4f),",
av_q2d(params->window_upper_left_corner_x),
av_q2d(params->window_upper_left_corner_y));
av_log(ctx, AV_LOG_INFO, "window_lower_right_corner: (%5.4f,%5.4f), ",
av_q2d(params->window_lower_right_corner_x),
av_q2d(params->window_lower_right_corner_y));
av_log(ctx, AV_LOG_INFO, "window_upper_left_corner: (%5.4f, %5.4f), ",
av_q2d(params->window_upper_left_corner_x),
av_q2d(params->window_upper_left_corner_y));
av_log(ctx, AV_LOG_INFO, "center_of_ellipse_x: (%d,%d), ",
params->center_of_ellipse_x,
params->center_of_ellipse_y);
av_log(ctx, AV_LOG_INFO, "rotation_angle: %d, ",
params->rotation_angle);
av_log(ctx, AV_LOG_INFO, "semimajor_axis_internal_ellipse: %d, ",
params->semimajor_axis_internal_ellipse);
av_log(ctx, AV_LOG_INFO, "semimajor_axis_external_ellipse: %d, ",
params->semimajor_axis_external_ellipse);
av_log(ctx, AV_LOG_INFO, "semiminor_axis_external_ellipse: %d, ",
params->semiminor_axis_external_ellipse);
av_log(ctx, AV_LOG_INFO, "overlap_process_option: %d}, ",
params->overlap_process_option);
}
av_log(ctx, AV_LOG_INFO, "targeted_system_display_maximum_luminance: %9.4f, ",
av_q2d(hdr_plus->targeted_system_display_maximum_luminance));
if (hdr_plus->targeted_system_display_actual_peak_luminance_flag) {
av_log(ctx, AV_LOG_INFO, "targeted_system_display_actual_peak_luminance: {");
for (int i = 0; i < hdr_plus->num_rows_targeted_system_display_actual_peak_luminance; i++) {
av_log(ctx, AV_LOG_INFO, "(");
for (int j = 0; j < hdr_plus->num_cols_targeted_system_display_actual_peak_luminance; j++) {
av_log(ctx, AV_LOG_INFO, "%5.4f,",
av_q2d(hdr_plus->targeted_system_display_actual_peak_luminance[i][j]));
}
av_log(ctx, AV_LOG_INFO, ")");
}
av_log(ctx, AV_LOG_INFO, "}, ");
}
for (int w = 0; w < hdr_plus->num_windows; w++) {
AVHDRPlusColorTransformParams *params = &hdr_plus->params[w];
av_log(ctx, AV_LOG_INFO, "window %d {maxscl: {", w);
for (int i = 0; i < 3; i++) {
av_log(ctx, AV_LOG_INFO, "%5.4f,",av_q2d(params->maxscl[i]));
}
av_log(ctx, AV_LOG_INFO, "} average_maxrgb: %5.4f, ",
av_q2d(params->average_maxrgb));
av_log(ctx, AV_LOG_INFO, "distribution_maxrgb: {");
for (int i = 0; i < params->num_distribution_maxrgb_percentiles; i++) {
av_log(ctx, AV_LOG_INFO, "(%d,%5.4f)",
params->distribution_maxrgb[i].percentage,
av_q2d(params->distribution_maxrgb[i].percentile));
}
av_log(ctx, AV_LOG_INFO, "} fraction_bright_pixels: %5.4f, ",
av_q2d(params->fraction_bright_pixels));
if (params->tone_mapping_flag) {
av_log(ctx, AV_LOG_INFO, "knee_point: (%5.4f,%5.4f), ", av_q2d(params->knee_point_x), av_q2d(params->knee_point_y));
av_log(ctx, AV_LOG_INFO, "bezier_curve_anchors: {");
for (int i = 0; i < params->num_bezier_curve_anchors; i++) {
av_log(ctx, AV_LOG_INFO, "%5.4f,",
av_q2d(params->bezier_curve_anchors[i]));
}
av_log(ctx, AV_LOG_INFO, "} ");
}
if (params->color_saturation_mapping_flag) {
av_log(ctx, AV_LOG_INFO, "color_saturation_weight: %5.4f",
av_q2d(params->color_saturation_weight));
}
av_log(ctx, AV_LOG_INFO, "} ");
}
if (hdr_plus->mastering_display_actual_peak_luminance_flag) {
av_log(ctx, AV_LOG_INFO, "mastering_display_actual_peak_luminance: {");
for (int i = 0; i < hdr_plus->num_rows_mastering_display_actual_peak_luminance; i++) {
av_log(ctx, AV_LOG_INFO, "(");
for (int j = 0; j < hdr_plus->num_cols_mastering_display_actual_peak_luminance; j++) {
av_log(ctx, AV_LOG_INFO, " %5.4f,",
av_q2d(hdr_plus->mastering_display_actual_peak_luminance[i][j]));
}
av_log(ctx, AV_LOG_INFO, ")");
}
av_log(ctx, AV_LOG_INFO, "} ");
}
}
static void dump_content_light_metadata(AVFilterContext *ctx, AVFrameSideData *sd)
{
const AVContentLightMetadata *metadata = (const AVContentLightMetadata *)sd->data;
av_log(ctx, AV_LOG_INFO, "Content Light Level information: "
"MaxCLL=%d, MaxFALL=%d",
metadata->MaxCLL, metadata->MaxFALL);
}
static void dump_video_enc_params(AVFilterContext *ctx, const AVFrameSideData *sd)
{
const AVVideoEncParams *par = (const AVVideoEncParams *)sd->data;
int plane, acdc;
av_log(ctx, AV_LOG_INFO, "video encoding parameters: type %d; ", par->type);
if (par->qp)
av_log(ctx, AV_LOG_INFO, "qp=%d; ", par->qp);
for (plane = 0; plane < FF_ARRAY_ELEMS(par->delta_qp); plane++)
for (acdc = 0; acdc < FF_ARRAY_ELEMS(par->delta_qp[plane]); acdc++) {
int delta_qp = par->delta_qp[plane][acdc];
if (delta_qp)
av_log(ctx, AV_LOG_INFO, "delta_qp[%d][%d]=%d; ",
plane, acdc, delta_qp);
}
if (par->nb_blocks)
av_log(ctx, AV_LOG_INFO, "%u blocks; ", par->nb_blocks);
}
static void dump_sei_unregistered_metadata(AVFilterContext *ctx, const AVFrameSideData *sd)
{
const int uuid_size = 16;
const uint8_t *user_data = sd->data;
int i;
if (sd->size < uuid_size) {
av_log(ctx, AV_LOG_ERROR, "invalid data(%"SIZE_SPECIFIER" < "
"UUID(%d-bytes))\n", sd->size, uuid_size);
return;
}
av_log(ctx, AV_LOG_INFO, "User Data Unregistered:\n");
av_log(ctx, AV_LOG_INFO, "UUID=");
for (i = 0; i < uuid_size; i++) {
av_log(ctx, AV_LOG_INFO, "%02x", user_data[i]);
if (i == 3 || i == 5 || i == 7 || i == 9)
av_log(ctx, AV_LOG_INFO, "-");
}
av_log(ctx, AV_LOG_INFO, "\n");
av_log(ctx, AV_LOG_INFO, "User Data=");
for (; i < sd->size; i++) {
av_log(ctx, AV_LOG_INFO, "%02x", user_data[i]);
}
av_log(ctx, AV_LOG_INFO, "\n");
}
static void dump_sei_film_grain_params_metadata(AVFilterContext *ctx, const AVFrameSideData *sd)
{
const AVFilmGrainParams *fgp = (const AVFilmGrainParams *)sd->data;
const char *const film_grain_type_names[] = {
[AV_FILM_GRAIN_PARAMS_NONE] = "none",
[AV_FILM_GRAIN_PARAMS_AV1] = "av1",
[AV_FILM_GRAIN_PARAMS_H274] = "h274",
};
if (fgp->type >= FF_ARRAY_ELEMS(film_grain_type_names)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
av_log(ctx, AV_LOG_INFO, "film grain parameters: type %s; ", film_grain_type_names[fgp->type]);
av_log(ctx, AV_LOG_INFO, "seed=%"PRIu64"; ", fgp->seed);
switch (fgp->type) {
case AV_FILM_GRAIN_PARAMS_NONE:
case AV_FILM_GRAIN_PARAMS_AV1:
return;
case AV_FILM_GRAIN_PARAMS_H274: {
const AVFilmGrainH274Params *h274 = &fgp->codec.h274;
const char *color_range_str = av_color_range_name(h274->color_range);
const char *color_primaries_str = av_color_primaries_name(h274->color_primaries);
const char *color_trc_str = av_color_transfer_name(h274->color_trc);
const char *colorspace_str = av_color_space_name(h274->color_space);
av_log(ctx, AV_LOG_INFO, "model_id=%d; ", h274->model_id);
av_log(ctx, AV_LOG_INFO, "bit_depth_luma=%d; ", h274->bit_depth_luma);
av_log(ctx, AV_LOG_INFO, "bit_depth_chroma=%d; ", h274->bit_depth_chroma);
av_log(ctx, AV_LOG_INFO, "color_range=%s; ", color_range_str ? color_range_str : "unknown");
av_log(ctx, AV_LOG_INFO, "color_primaries=%s; ", color_primaries_str ? color_primaries_str : "unknown");
av_log(ctx, AV_LOG_INFO, "color_trc=%s; ", color_trc_str ? color_trc_str : "unknown");
av_log(ctx, AV_LOG_INFO, "color_space=%s; ", colorspace_str ? colorspace_str : "unknown");
av_log(ctx, AV_LOG_INFO, "blending_mode_id=%d; ", h274->blending_mode_id);
av_log(ctx, AV_LOG_INFO, "log2_scale_factor=%d; ", h274->log2_scale_factor);
for (int c = 0; c < 3; c++)
if (h274->component_model_present[c] && (h274->num_model_values[c] > 6 ||
h274->num_intensity_intervals[c] < 1 ||
h274->num_intensity_intervals[c] > 256)) {
av_log(ctx, AV_LOG_ERROR, "invalid data\n");
return;
}
for (int c = 0; c < 3; c++) {
if (!h274->component_model_present[c])
continue;
av_log(ctx, AV_LOG_INFO, "num_intensity_intervals[%d]=%u; ", c, h274->num_intensity_intervals[c]);
av_log(ctx, AV_LOG_INFO, "num_model_values[%d]=%u; ", c, h274->num_model_values[c]);
for (int i = 0; i < h274->num_intensity_intervals[c]; i++) {
av_log(ctx, AV_LOG_INFO, "intensity_interval_lower_bound[%d][%d]=%u; ",
c, i, h274->intensity_interval_lower_bound[c][i]);
av_log(ctx, AV_LOG_INFO, "intensity_interval_upper_bound[%d][%d]=%u; ",
c, i, h274->intensity_interval_upper_bound[c][i]);
for (int j = 0; j < h274->num_model_values[c]; j++)
av_log(ctx, AV_LOG_INFO, "comp_model_value[%d][%d][%d]=%d; ",
c, i, j, h274->comp_model_value[c][i][j]);
}
}
break;
}
}
}
static void dump_color_property(AVFilterContext *ctx, AVFrame *frame)
{
const char *color_range_str = av_color_range_name(frame->color_range);
const char *colorspace_str = av_color_space_name(frame->colorspace);
const char *color_primaries_str = av_color_primaries_name(frame->color_primaries);
const char *color_trc_str = av_color_transfer_name(frame->color_trc);
if (!color_range_str || frame->color_range == AVCOL_RANGE_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, "color_range:unknown");
} else {
av_log(ctx, AV_LOG_INFO, "color_range:%s", color_range_str);
}
if (!colorspace_str || frame->colorspace == AVCOL_SPC_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_space:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_space:%s", colorspace_str);
}
if (!color_primaries_str || frame->color_primaries == AVCOL_PRI_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_primaries:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_primaries:%s", color_primaries_str);
}
if (!color_trc_str || frame->color_trc == AVCOL_TRC_UNSPECIFIED) {
av_log(ctx, AV_LOG_INFO, " color_trc:unknown");
} else {
av_log(ctx, AV_LOG_INFO, " color_trc:%s", color_trc_str);
}
av_log(ctx, AV_LOG_INFO, "\n");
}
static void update_sample_stats_8(const uint8_t *src, int len, int64_t *sum, int64_t *sum2)
{
int i;
for (i = 0; i < len; i++) {
*sum += src[i];
*sum2 += src[i] * src[i];
}
}
static void update_sample_stats_16(int be, const uint8_t *src, int len, int64_t *sum, int64_t *sum2)
{
const uint16_t *src1 = (const uint16_t *)src;
int i;
for (i = 0; i < len / 2; i++) {
if ((HAVE_BIGENDIAN && !be) || (!HAVE_BIGENDIAN && be)) {
*sum += av_bswap16(src1[i]);
*sum2 += (uint32_t)av_bswap16(src1[i]) * (uint32_t)av_bswap16(src1[i]);
} else {
*sum += src1[i];
*sum2 += (uint32_t)src1[i] * (uint32_t)src1[i];
}
}
}
static void update_sample_stats(int depth, int be, const uint8_t *src, int len, int64_t *sum, int64_t *sum2)
{
if (depth <= 8)
update_sample_stats_8(src, len, sum, sum2);
else
update_sample_stats_16(be, src, len, sum, sum2);
}
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
{
AVFilterContext *ctx = inlink->dst;
ShowInfoContext *s = ctx->priv;
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
uint32_t plane_checksum[4] = {0}, checksum = 0;
int64_t sum[4] = {0}, sum2[4] = {0};
int32_t pixelcount[4] = {0};
int bitdepth = desc->comp[0].depth;
int be = desc->flags & AV_PIX_FMT_FLAG_BE;
int i, plane, vsub = desc->log2_chroma_h;
for (plane = 0; plane < 4 && s->calculate_checksums && frame->data[plane] && frame->linesize[plane]; plane++) {
uint8_t *data = frame->data[plane];
int h = plane == 1 || plane == 2 ? AV_CEIL_RSHIFT(inlink->h, vsub) : inlink->h;
int linesize = av_image_get_linesize(frame->format, frame->width, plane);
int width = linesize >> (bitdepth > 8);
if (linesize < 0)
return linesize;
for (i = 0; i < h; i++) {
plane_checksum[plane] = av_adler32_update(plane_checksum[plane], data, linesize);
checksum = av_adler32_update(checksum, data, linesize);
update_sample_stats(bitdepth, be, data, linesize, sum+plane, sum2+plane);
pixelcount[plane] += width;
data += frame->linesize[plane];
}
}
av_log(ctx, AV_LOG_INFO,
"n:%4"PRId64" pts:%7s pts_time:%-7s pos:%9"PRId64" "
"fmt:%s sar:%d/%d s:%dx%d i:%c iskey:%d type:%c ",
inlink->frame_count_out,
av_ts2str(frame->pts), av_ts2timestr(frame->pts, &inlink->time_base), frame->pkt_pos,
desc->name,
frame->sample_aspect_ratio.num, frame->sample_aspect_ratio.den,
frame->width, frame->height,
!frame->interlaced_frame ? 'P' : /* Progressive */
frame->top_field_first ? 'T' : 'B', /* Top / Bottom */
frame->key_frame,
av_get_picture_type_char(frame->pict_type));
if (s->calculate_checksums) {
av_log(ctx, AV_LOG_INFO,
"checksum:%08"PRIX32" plane_checksum:[%08"PRIX32,
checksum, plane_checksum[0]);
for (plane = 1; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, " %08"PRIX32, plane_checksum[plane]);
av_log(ctx, AV_LOG_INFO, "] mean:[");
for (plane = 0; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, "%"PRId64" ", (sum[plane] + pixelcount[plane]/2) / pixelcount[plane]);
av_log(ctx, AV_LOG_INFO, "\b] stdev:[");
for (plane = 0; plane < 4 && frame->data[plane] && frame->linesize[plane]; plane++)
av_log(ctx, AV_LOG_INFO, "%3.1f ",
sqrt((sum2[plane] - sum[plane]*(double)sum[plane]/pixelcount[plane])/pixelcount[plane]));
av_log(ctx, AV_LOG_INFO, "\b]");
}
av_log(ctx, AV_LOG_INFO, "\n");
for (i = 0; i < frame->nb_side_data; i++) {
AVFrameSideData *sd = frame->side_data[i];
av_log(ctx, AV_LOG_INFO, " side data - ");
switch (sd->type) {
case AV_FRAME_DATA_PANSCAN:
av_log(ctx, AV_LOG_INFO, "pan/scan");
break;
case AV_FRAME_DATA_A53_CC:
av_log(ctx, AV_LOG_INFO, "A/53 closed captions "
"(%"SIZE_SPECIFIER" bytes)", sd->size);
break;
case AV_FRAME_DATA_SPHERICAL:
dump_spherical(ctx, frame, sd);
break;
case AV_FRAME_DATA_STEREO3D:
dump_stereo3d(ctx, sd);
break;
case AV_FRAME_DATA_S12M_TIMECODE: {
dump_s12m_timecode(ctx, inlink->frame_rate, sd);
break;
}
case AV_FRAME_DATA_DISPLAYMATRIX:
av_log(ctx, AV_LOG_INFO, "displaymatrix: rotation of %.2f degrees",
av_display_rotation_get((int32_t *)sd->data));
break;
case AV_FRAME_DATA_AFD:
av_log(ctx, AV_LOG_INFO, "afd: value of %"PRIu8, sd->data[0]);
break;
case AV_FRAME_DATA_REGIONS_OF_INTEREST:
dump_roi(ctx, sd);
break;
case AV_FRAME_DATA_DETECTION_BBOXES:
dump_detection_bbox(ctx, sd);
break;
case AV_FRAME_DATA_MASTERING_DISPLAY_METADATA:
dump_mastering_display(ctx, sd);
break;
case AV_FRAME_DATA_DYNAMIC_HDR_PLUS:
dump_dynamic_hdr_plus(ctx, sd);
break;
case AV_FRAME_DATA_CONTENT_LIGHT_LEVEL:
dump_content_light_metadata(ctx, sd);
break;
case AV_FRAME_DATA_GOP_TIMECODE: {
char tcbuf[AV_TIMECODE_STR_SIZE];
av_timecode_make_mpeg_tc_string(tcbuf, *(int64_t *)(sd->data));
av_log(ctx, AV_LOG_INFO, "GOP timecode - %s", tcbuf);
break;
}
case AV_FRAME_DATA_VIDEO_ENC_PARAMS:
dump_video_enc_params(ctx, sd);
break;
case AV_FRAME_DATA_SEI_UNREGISTERED:
dump_sei_unregistered_metadata(ctx, sd);
break;
case AV_FRAME_DATA_FILM_GRAIN_PARAMS:
dump_sei_film_grain_params_metadata(ctx, sd);
break;
default:
av_log(ctx, AV_LOG_WARNING, "unknown side data type %d "
"(%"SIZE_SPECIFIER" bytes)\n", sd->type, sd->size);
break;
}
av_log(ctx, AV_LOG_INFO, "\n");
}
dump_color_property(ctx, frame);
return ff_filter_frame(inlink->dst->outputs[0], frame);
}
static int config_props(AVFilterContext *ctx, AVFilterLink *link, int is_out)
{
av_log(ctx, AV_LOG_INFO, "config %s time_base: %d/%d, frame_rate: %d/%d\n",
is_out ? "out" : "in",
link->time_base.num, link->time_base.den,
link->frame_rate.num, link->frame_rate.den);
return 0;
}
static int config_props_in(AVFilterLink *link)
{
AVFilterContext *ctx = link->dst;
return config_props(ctx, link, 0);
}
static int config_props_out(AVFilterLink *link)
{
AVFilterContext *ctx = link->src;
return config_props(ctx, link, 1);
}
static const AVFilterPad avfilter_vf_showinfo_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = filter_frame,
.config_props = config_props_in,
},
};
static const AVFilterPad avfilter_vf_showinfo_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_props_out,
},
};
const AVFilter ff_vf_showinfo = {
.name = "showinfo",
.description = NULL_IF_CONFIG_SMALL("Show textual information for each video frame."),
FILTER_INPUTS(avfilter_vf_showinfo_inputs),
FILTER_OUTPUTS(avfilter_vf_showinfo_outputs),
.priv_size = sizeof(ShowInfoContext),
.priv_class = &showinfo_class,
};