mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
a247ac640d
Given that the AVCodec.next pointer has now been removed, most of the AVCodecs are not modified at all any more and can therefore be made const (as this patch does); the only exceptions are the very few codecs for external libraries that have a init_static_data callback. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@gmail.com> Signed-off-by: James Almer <jamrial@gmail.com>
1159 lines
36 KiB
C
1159 lines
36 KiB
C
/*
|
|
* AV1 video decoder
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/film_grain_params.h"
|
|
#include "libavutil/pixdesc.h"
|
|
#include "libavutil/opt.h"
|
|
#include "avcodec.h"
|
|
#include "av1dec.h"
|
|
#include "bytestream.h"
|
|
#include "hwconfig.h"
|
|
#include "internal.h"
|
|
#include "profiles.h"
|
|
|
|
static uint32_t inverse_recenter(int r, uint32_t v)
|
|
{
|
|
if (v > 2 * r)
|
|
return v;
|
|
else if (v & 1)
|
|
return r - ((v + 1) >> 1);
|
|
else
|
|
return r + (v >> 1);
|
|
}
|
|
|
|
static uint32_t decode_unsigned_subexp_with_ref(uint32_t sub_exp,
|
|
int mx, int r)
|
|
{
|
|
if ((r << 1) <= mx) {
|
|
return inverse_recenter(r, sub_exp);
|
|
} else {
|
|
return mx - 1 - inverse_recenter(mx - 1 - r, sub_exp);
|
|
}
|
|
}
|
|
|
|
static int32_t decode_signed_subexp_with_ref(uint32_t sub_exp, int low,
|
|
int high, int r)
|
|
{
|
|
int32_t x = decode_unsigned_subexp_with_ref(sub_exp, high - low, r - low);
|
|
return x + low;
|
|
}
|
|
|
|
static void read_global_param(AV1DecContext *s, int type, int ref, int idx)
|
|
{
|
|
uint8_t primary_frame, prev_frame;
|
|
uint32_t abs_bits, prec_bits, round, prec_diff, sub, mx;
|
|
int32_t r, prev_gm_param;
|
|
|
|
primary_frame = s->raw_frame_header->primary_ref_frame;
|
|
prev_frame = s->raw_frame_header->ref_frame_idx[primary_frame];
|
|
abs_bits = AV1_GM_ABS_ALPHA_BITS;
|
|
prec_bits = AV1_GM_ALPHA_PREC_BITS;
|
|
|
|
/* setup_past_independence() sets PrevGmParams to default values. We can
|
|
* simply point to the current's frame gm_params as they will be initialized
|
|
* with defaults at this point.
|
|
*/
|
|
if (s->raw_frame_header->primary_ref_frame == AV1_PRIMARY_REF_NONE)
|
|
prev_gm_param = s->cur_frame.gm_params[ref][idx];
|
|
else
|
|
prev_gm_param = s->ref[prev_frame].gm_params[ref][idx];
|
|
|
|
if (idx < 2) {
|
|
if (type == AV1_WARP_MODEL_TRANSLATION) {
|
|
abs_bits = AV1_GM_ABS_TRANS_ONLY_BITS -
|
|
!s->raw_frame_header->allow_high_precision_mv;
|
|
prec_bits = AV1_GM_TRANS_ONLY_PREC_BITS -
|
|
!s->raw_frame_header->allow_high_precision_mv;
|
|
} else {
|
|
abs_bits = AV1_GM_ABS_TRANS_BITS;
|
|
prec_bits = AV1_GM_TRANS_PREC_BITS;
|
|
}
|
|
}
|
|
round = (idx % 3) == 2 ? (1 << AV1_WARPEDMODEL_PREC_BITS) : 0;
|
|
prec_diff = AV1_WARPEDMODEL_PREC_BITS - prec_bits;
|
|
sub = (idx % 3) == 2 ? (1 << prec_bits) : 0;
|
|
mx = 1 << abs_bits;
|
|
r = (prev_gm_param >> prec_diff) - sub;
|
|
|
|
s->cur_frame.gm_params[ref][idx] =
|
|
(decode_signed_subexp_with_ref(s->raw_frame_header->gm_params[ref][idx],
|
|
-mx, mx + 1, r) << prec_diff) + round;
|
|
}
|
|
|
|
/**
|
|
* update gm type/params, since cbs already implemented part of this funcation,
|
|
* so we don't need to full implement spec.
|
|
*/
|
|
static void global_motion_params(AV1DecContext *s)
|
|
{
|
|
const AV1RawFrameHeader *header = s->raw_frame_header;
|
|
int type, ref;
|
|
|
|
for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
|
|
s->cur_frame.gm_type[ref] = AV1_WARP_MODEL_IDENTITY;
|
|
for (int i = 0; i < 6; i++)
|
|
s->cur_frame.gm_params[ref][i] = (i % 3 == 2) ?
|
|
1 << AV1_WARPEDMODEL_PREC_BITS : 0;
|
|
}
|
|
if (header->frame_type == AV1_FRAME_KEY ||
|
|
header->frame_type == AV1_FRAME_INTRA_ONLY)
|
|
return;
|
|
|
|
for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
|
|
if (header->is_global[ref]) {
|
|
if (header->is_rot_zoom[ref]) {
|
|
type = AV1_WARP_MODEL_ROTZOOM;
|
|
} else {
|
|
type = header->is_translation[ref] ? AV1_WARP_MODEL_TRANSLATION
|
|
: AV1_WARP_MODEL_AFFINE;
|
|
}
|
|
} else {
|
|
type = AV1_WARP_MODEL_IDENTITY;
|
|
}
|
|
s->cur_frame.gm_type[ref] = type;
|
|
|
|
if (type >= AV1_WARP_MODEL_ROTZOOM) {
|
|
read_global_param(s, type, ref, 2);
|
|
read_global_param(s, type, ref, 3);
|
|
if (type == AV1_WARP_MODEL_AFFINE) {
|
|
read_global_param(s, type, ref, 4);
|
|
read_global_param(s, type, ref, 5);
|
|
} else {
|
|
s->cur_frame.gm_params[ref][4] = -s->cur_frame.gm_params[ref][3];
|
|
s->cur_frame.gm_params[ref][5] = s->cur_frame.gm_params[ref][2];
|
|
}
|
|
}
|
|
if (type >= AV1_WARP_MODEL_TRANSLATION) {
|
|
read_global_param(s, type, ref, 0);
|
|
read_global_param(s, type, ref, 1);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int get_relative_dist(const AV1RawSequenceHeader *seq,
|
|
unsigned int a, unsigned int b)
|
|
{
|
|
unsigned int diff = a - b;
|
|
unsigned int m = 1 << seq->order_hint_bits_minus_1;
|
|
return (diff & (m - 1)) - (diff & m);
|
|
}
|
|
|
|
static void skip_mode_params(AV1DecContext *s)
|
|
{
|
|
const AV1RawFrameHeader *header = s->raw_frame_header;
|
|
const AV1RawSequenceHeader *seq = s->raw_seq;
|
|
|
|
int forward_idx, backward_idx;
|
|
int forward_hint, backward_hint;
|
|
int second_forward_idx, second_forward_hint;
|
|
int ref_hint, dist, i;
|
|
|
|
if (!header->skip_mode_present)
|
|
return;
|
|
|
|
forward_idx = -1;
|
|
backward_idx = -1;
|
|
for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
|
|
ref_hint = s->ref[header->ref_frame_idx[i]].raw_frame_header->order_hint;
|
|
dist = get_relative_dist(seq, ref_hint, header->order_hint);
|
|
if (dist < 0) {
|
|
if (forward_idx < 0 ||
|
|
get_relative_dist(seq, ref_hint, forward_hint) > 0) {
|
|
forward_idx = i;
|
|
forward_hint = ref_hint;
|
|
}
|
|
} else if (dist > 0) {
|
|
if (backward_idx < 0 ||
|
|
get_relative_dist(seq, ref_hint, backward_hint) < 0) {
|
|
backward_idx = i;
|
|
backward_hint = ref_hint;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (forward_idx < 0) {
|
|
return;
|
|
} else if (backward_idx >= 0) {
|
|
s->cur_frame.skip_mode_frame_idx[0] =
|
|
AV1_REF_FRAME_LAST + FFMIN(forward_idx, backward_idx);
|
|
s->cur_frame.skip_mode_frame_idx[1] =
|
|
AV1_REF_FRAME_LAST + FFMAX(forward_idx, backward_idx);
|
|
return;
|
|
}
|
|
|
|
second_forward_idx = -1;
|
|
for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
|
|
ref_hint = s->ref[header->ref_frame_idx[i]].raw_frame_header->order_hint;
|
|
if (get_relative_dist(seq, ref_hint, forward_hint) < 0) {
|
|
if (second_forward_idx < 0 ||
|
|
get_relative_dist(seq, ref_hint, second_forward_hint) > 0) {
|
|
second_forward_idx = i;
|
|
second_forward_hint = ref_hint;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (second_forward_idx < 0)
|
|
return;
|
|
|
|
s->cur_frame.skip_mode_frame_idx[0] =
|
|
AV1_REF_FRAME_LAST + FFMIN(forward_idx, second_forward_idx);
|
|
s->cur_frame.skip_mode_frame_idx[1] =
|
|
AV1_REF_FRAME_LAST + FFMAX(forward_idx, second_forward_idx);
|
|
}
|
|
|
|
static void coded_lossless_param(AV1DecContext *s)
|
|
{
|
|
const AV1RawFrameHeader *header = s->raw_frame_header;
|
|
int i;
|
|
|
|
if (header->delta_q_y_dc || header->delta_q_u_ac ||
|
|
header->delta_q_u_dc || header->delta_q_v_ac ||
|
|
header->delta_q_v_dc) {
|
|
s->cur_frame.coded_lossless = 0;
|
|
return;
|
|
}
|
|
|
|
s->cur_frame.coded_lossless = 1;
|
|
for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
|
|
int qindex;
|
|
if (header->feature_enabled[i][AV1_SEG_LVL_ALT_Q]) {
|
|
qindex = (header->base_q_idx +
|
|
header->feature_value[i][AV1_SEG_LVL_ALT_Q]);
|
|
} else {
|
|
qindex = header->base_q_idx;
|
|
}
|
|
qindex = av_clip_uintp2(qindex, 8);
|
|
|
|
if (qindex) {
|
|
s->cur_frame.coded_lossless = 0;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void load_grain_params(AV1DecContext *s)
|
|
{
|
|
const AV1RawFrameHeader *header = s->raw_frame_header;
|
|
const AV1RawFilmGrainParams *film_grain = &header->film_grain, *src;
|
|
AV1RawFilmGrainParams *dst = &s->cur_frame.film_grain;
|
|
|
|
if (!film_grain->apply_grain)
|
|
return;
|
|
|
|
if (film_grain->update_grain) {
|
|
memcpy(dst, film_grain, sizeof(*dst));
|
|
return;
|
|
}
|
|
|
|
src = &s->ref[film_grain->film_grain_params_ref_idx].film_grain;
|
|
|
|
memcpy(dst, src, sizeof(*dst));
|
|
dst->grain_seed = film_grain->grain_seed;
|
|
}
|
|
|
|
static int init_tile_data(AV1DecContext *s)
|
|
|
|
{
|
|
int cur_tile_num =
|
|
s->raw_frame_header->tile_cols * s->raw_frame_header->tile_rows;
|
|
if (s->tile_num < cur_tile_num) {
|
|
int ret = av_reallocp_array(&s->tile_group_info, cur_tile_num,
|
|
sizeof(TileGroupInfo));
|
|
if (ret < 0) {
|
|
s->tile_num = 0;
|
|
return ret;
|
|
}
|
|
}
|
|
s->tile_num = cur_tile_num;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int get_tiles_info(AVCodecContext *avctx, const AV1RawTileGroup *tile_group)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
GetByteContext gb;
|
|
uint16_t tile_num, tile_row, tile_col;
|
|
uint32_t size = 0, size_bytes = 0;
|
|
|
|
bytestream2_init(&gb, tile_group->tile_data.data,
|
|
tile_group->tile_data.data_size);
|
|
s->tg_start = tile_group->tg_start;
|
|
s->tg_end = tile_group->tg_end;
|
|
|
|
for (tile_num = tile_group->tg_start; tile_num <= tile_group->tg_end; tile_num++) {
|
|
tile_row = tile_num / s->raw_frame_header->tile_cols;
|
|
tile_col = tile_num % s->raw_frame_header->tile_cols;
|
|
|
|
if (tile_num == tile_group->tg_end) {
|
|
s->tile_group_info[tile_num].tile_size = bytestream2_get_bytes_left(&gb);
|
|
s->tile_group_info[tile_num].tile_offset = bytestream2_tell(&gb);
|
|
s->tile_group_info[tile_num].tile_row = tile_row;
|
|
s->tile_group_info[tile_num].tile_column = tile_col;
|
|
return 0;
|
|
}
|
|
size_bytes = s->raw_frame_header->tile_size_bytes_minus1 + 1;
|
|
if (bytestream2_get_bytes_left(&gb) < size_bytes)
|
|
return AVERROR_INVALIDDATA;
|
|
size = 0;
|
|
for (int i = 0; i < size_bytes; i++)
|
|
size |= bytestream2_get_byteu(&gb) << 8 * i;
|
|
if (bytestream2_get_bytes_left(&gb) <= size)
|
|
return AVERROR_INVALIDDATA;
|
|
size++;
|
|
|
|
s->tile_group_info[tile_num].tile_size = size;
|
|
s->tile_group_info[tile_num].tile_offset = bytestream2_tell(&gb);
|
|
s->tile_group_info[tile_num].tile_row = tile_row;
|
|
s->tile_group_info[tile_num].tile_column = tile_col;
|
|
|
|
bytestream2_skipu(&gb, size);
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int get_pixel_format(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
const AV1RawSequenceHeader *seq = s->raw_seq;
|
|
uint8_t bit_depth;
|
|
int ret;
|
|
enum AVPixelFormat pix_fmt = AV_PIX_FMT_NONE;
|
|
#define HWACCEL_MAX (CONFIG_AV1_DXVA2_HWACCEL + \
|
|
CONFIG_AV1_D3D11VA_HWACCEL * 2 + \
|
|
CONFIG_AV1_NVDEC_HWACCEL + \
|
|
CONFIG_AV1_VAAPI_HWACCEL)
|
|
enum AVPixelFormat pix_fmts[HWACCEL_MAX + 2], *fmtp = pix_fmts;
|
|
|
|
if (seq->seq_profile == 2 && seq->color_config.high_bitdepth)
|
|
bit_depth = seq->color_config.twelve_bit ? 12 : 10;
|
|
else if (seq->seq_profile <= 2)
|
|
bit_depth = seq->color_config.high_bitdepth ? 10 : 8;
|
|
else {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Unknown AV1 profile %d.\n", seq->seq_profile);
|
|
return -1;
|
|
}
|
|
|
|
if (!seq->color_config.mono_chrome) {
|
|
// 4:4:4 x:0 y:0, 4:2:2 x:1 y:0, 4:2:0 x:1 y:1
|
|
if (seq->color_config.subsampling_x == 0 &&
|
|
seq->color_config.subsampling_y == 0) {
|
|
if (bit_depth == 8)
|
|
pix_fmt = AV_PIX_FMT_YUV444P;
|
|
else if (bit_depth == 10)
|
|
pix_fmt = AV_PIX_FMT_YUV444P10;
|
|
else if (bit_depth == 12)
|
|
pix_fmt = AV_PIX_FMT_YUV444P12;
|
|
else
|
|
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
|
|
} else if (seq->color_config.subsampling_x == 1 &&
|
|
seq->color_config.subsampling_y == 0) {
|
|
if (bit_depth == 8)
|
|
pix_fmt = AV_PIX_FMT_YUV422P;
|
|
else if (bit_depth == 10)
|
|
pix_fmt = AV_PIX_FMT_YUV422P10;
|
|
else if (bit_depth == 12)
|
|
pix_fmt = AV_PIX_FMT_YUV422P12;
|
|
else
|
|
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
|
|
} else if (seq->color_config.subsampling_x == 1 &&
|
|
seq->color_config.subsampling_y == 1) {
|
|
if (bit_depth == 8)
|
|
pix_fmt = AV_PIX_FMT_YUV420P;
|
|
else if (bit_depth == 10)
|
|
pix_fmt = AV_PIX_FMT_YUV420P10;
|
|
else if (bit_depth == 12)
|
|
pix_fmt = AV_PIX_FMT_YUV420P12;
|
|
else
|
|
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
|
|
}
|
|
} else {
|
|
if (bit_depth == 8)
|
|
pix_fmt = AV_PIX_FMT_GRAY8;
|
|
else if (bit_depth == 10)
|
|
pix_fmt = AV_PIX_FMT_GRAY10;
|
|
else if (bit_depth == 12)
|
|
pix_fmt = AV_PIX_FMT_GRAY12;
|
|
else
|
|
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_DEBUG, "AV1 decode get format: %s.\n",
|
|
av_get_pix_fmt_name(pix_fmt));
|
|
|
|
if (pix_fmt == AV_PIX_FMT_NONE)
|
|
return -1;
|
|
s->pix_fmt = pix_fmt;
|
|
|
|
switch (s->pix_fmt) {
|
|
case AV_PIX_FMT_YUV420P:
|
|
#if CONFIG_AV1_DXVA2_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_DXVA2_VLD;
|
|
#endif
|
|
#if CONFIG_AV1_D3D11VA_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_D3D11VA_VLD;
|
|
*fmtp++ = AV_PIX_FMT_D3D11;
|
|
#endif
|
|
#if CONFIG_AV1_NVDEC_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_CUDA;
|
|
#endif
|
|
#if CONFIG_AV1_VAAPI_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_VAAPI;
|
|
#endif
|
|
break;
|
|
case AV_PIX_FMT_YUV420P10:
|
|
#if CONFIG_AV1_DXVA2_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_DXVA2_VLD;
|
|
#endif
|
|
#if CONFIG_AV1_D3D11VA_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_D3D11VA_VLD;
|
|
*fmtp++ = AV_PIX_FMT_D3D11;
|
|
#endif
|
|
#if CONFIG_AV1_NVDEC_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_CUDA;
|
|
#endif
|
|
#if CONFIG_AV1_VAAPI_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_VAAPI;
|
|
#endif
|
|
break;
|
|
case AV_PIX_FMT_GRAY8:
|
|
#if CONFIG_AV1_NVDEC_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_CUDA;
|
|
#endif
|
|
break;
|
|
case AV_PIX_FMT_GRAY10:
|
|
#if CONFIG_AV1_NVDEC_HWACCEL
|
|
*fmtp++ = AV_PIX_FMT_CUDA;
|
|
#endif
|
|
break;
|
|
}
|
|
|
|
*fmtp++ = s->pix_fmt;
|
|
*fmtp = AV_PIX_FMT_NONE;
|
|
|
|
ret = ff_thread_get_format(avctx, pix_fmts);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
/**
|
|
* check if the HW accel is inited correctly. If not, return un-implemented.
|
|
* Since now the av1 decoder doesn't support native decode, if it will be
|
|
* implemented in the future, need remove this check.
|
|
*/
|
|
if (!avctx->hwaccel) {
|
|
av_log(avctx, AV_LOG_ERROR, "Your platform doesn't suppport"
|
|
" hardware accelerated AV1 decoding.\n");
|
|
return AVERROR(ENOSYS);
|
|
}
|
|
|
|
avctx->pix_fmt = ret;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void av1_frame_unref(AVCodecContext *avctx, AV1Frame *f)
|
|
{
|
|
ff_thread_release_buffer(avctx, &f->tf);
|
|
av_buffer_unref(&f->hwaccel_priv_buf);
|
|
f->hwaccel_picture_private = NULL;
|
|
av_buffer_unref(&f->header_ref);
|
|
f->raw_frame_header = NULL;
|
|
f->spatial_id = f->temporal_id = 0;
|
|
memset(f->skip_mode_frame_idx, 0,
|
|
2 * sizeof(uint8_t));
|
|
memset(&f->film_grain, 0, sizeof(f->film_grain));
|
|
f->coded_lossless = 0;
|
|
}
|
|
|
|
static int av1_frame_ref(AVCodecContext *avctx, AV1Frame *dst, const AV1Frame *src)
|
|
{
|
|
int ret;
|
|
|
|
ret = ff_thread_ref_frame(&dst->tf, &src->tf);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
dst->header_ref = av_buffer_ref(src->header_ref);
|
|
if (!dst->header_ref)
|
|
goto fail;
|
|
|
|
dst->raw_frame_header = src->raw_frame_header;
|
|
|
|
if (src->hwaccel_picture_private) {
|
|
dst->hwaccel_priv_buf = av_buffer_ref(src->hwaccel_priv_buf);
|
|
if (!dst->hwaccel_priv_buf)
|
|
goto fail;
|
|
dst->hwaccel_picture_private = dst->hwaccel_priv_buf->data;
|
|
}
|
|
|
|
dst->spatial_id = src->spatial_id;
|
|
dst->temporal_id = src->temporal_id;
|
|
memcpy(dst->gm_type,
|
|
src->gm_type,
|
|
AV1_NUM_REF_FRAMES * sizeof(uint8_t));
|
|
memcpy(dst->gm_params,
|
|
src->gm_params,
|
|
AV1_NUM_REF_FRAMES * 6 * sizeof(int32_t));
|
|
memcpy(dst->skip_mode_frame_idx,
|
|
src->skip_mode_frame_idx,
|
|
2 * sizeof(uint8_t));
|
|
memcpy(&dst->film_grain,
|
|
&src->film_grain,
|
|
sizeof(dst->film_grain));
|
|
dst->coded_lossless = src->coded_lossless;
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
av1_frame_unref(avctx, dst);
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
static av_cold int av1_decode_free(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
|
|
av1_frame_unref(avctx, &s->ref[i]);
|
|
av_frame_free(&s->ref[i].tf.f);
|
|
}
|
|
av1_frame_unref(avctx, &s->cur_frame);
|
|
av_frame_free(&s->cur_frame.tf.f);
|
|
|
|
av_buffer_unref(&s->seq_ref);
|
|
av_buffer_unref(&s->header_ref);
|
|
av_freep(&s->tile_group_info);
|
|
|
|
ff_cbs_fragment_free(&s->current_obu);
|
|
ff_cbs_close(&s->cbc);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int set_context_with_sequence(AVCodecContext *avctx,
|
|
const AV1RawSequenceHeader *seq)
|
|
{
|
|
int width = seq->max_frame_width_minus_1 + 1;
|
|
int height = seq->max_frame_height_minus_1 + 1;
|
|
|
|
avctx->profile = seq->seq_profile;
|
|
avctx->level = seq->seq_level_idx[0];
|
|
|
|
avctx->color_range =
|
|
seq->color_config.color_range ? AVCOL_RANGE_JPEG : AVCOL_RANGE_MPEG;
|
|
avctx->color_primaries = seq->color_config.color_primaries;
|
|
avctx->colorspace = seq->color_config.color_primaries;
|
|
avctx->color_trc = seq->color_config.transfer_characteristics;
|
|
|
|
switch (seq->color_config.chroma_sample_position) {
|
|
case AV1_CSP_VERTICAL:
|
|
avctx->chroma_sample_location = AVCHROMA_LOC_LEFT;
|
|
break;
|
|
case AV1_CSP_COLOCATED:
|
|
avctx->chroma_sample_location = AVCHROMA_LOC_TOPLEFT;
|
|
break;
|
|
}
|
|
|
|
if (avctx->width != width || avctx->height != height) {
|
|
int ret = ff_set_dimensions(avctx, width, height);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
avctx->sample_aspect_ratio = (AVRational) { 1, 1 };
|
|
|
|
if (seq->timing_info.num_units_in_display_tick &&
|
|
seq->timing_info.time_scale) {
|
|
av_reduce(&avctx->framerate.den, &avctx->framerate.num,
|
|
seq->timing_info.num_units_in_display_tick,
|
|
seq->timing_info.time_scale,
|
|
INT_MAX);
|
|
if (seq->timing_info.equal_picture_interval)
|
|
avctx->ticks_per_frame = seq->timing_info.num_ticks_per_picture_minus_1 + 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int update_context_with_frame_header(AVCodecContext *avctx,
|
|
const AV1RawFrameHeader *header)
|
|
{
|
|
AVRational aspect_ratio;
|
|
int width = header->frame_width_minus_1 + 1;
|
|
int height = header->frame_height_minus_1 + 1;
|
|
int r_width = header->render_width_minus_1 + 1;
|
|
int r_height = header->render_height_minus_1 + 1;
|
|
int ret;
|
|
|
|
if (avctx->width != width || avctx->height != height) {
|
|
ret = ff_set_dimensions(avctx, width, height);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
av_reduce(&aspect_ratio.num, &aspect_ratio.den,
|
|
(int64_t)height * r_width,
|
|
(int64_t)width * r_height,
|
|
INT_MAX);
|
|
|
|
if (av_cmp_q(avctx->sample_aspect_ratio, aspect_ratio)) {
|
|
ret = ff_set_sar(avctx, aspect_ratio);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int av1_decode_init(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
AV1RawSequenceHeader *seq;
|
|
int ret;
|
|
|
|
s->avctx = avctx;
|
|
s->pix_fmt = AV_PIX_FMT_NONE;
|
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
|
|
s->ref[i].tf.f = av_frame_alloc();
|
|
if (!s->ref[i].tf.f) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Failed to allocate reference frame buffer %d.\n", i);
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
}
|
|
|
|
s->cur_frame.tf.f = av_frame_alloc();
|
|
if (!s->cur_frame.tf.f) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Failed to allocate current frame buffer.\n");
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
ret = ff_cbs_init(&s->cbc, AV_CODEC_ID_AV1, avctx);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
av_opt_set_int(s->cbc->priv_data, "operating_point", s->operating_point, 0);
|
|
|
|
if (avctx->extradata && avctx->extradata_size) {
|
|
ret = ff_cbs_read_extradata_from_codec(s->cbc,
|
|
&s->current_obu,
|
|
avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_WARNING, "Failed to read extradata.\n");
|
|
return ret;
|
|
}
|
|
|
|
seq = ((CodedBitstreamAV1Context *)(s->cbc->priv_data))->sequence_header;
|
|
if (!seq) {
|
|
av_log(avctx, AV_LOG_WARNING, "No sequence header available.\n");
|
|
goto end;
|
|
}
|
|
|
|
ret = set_context_with_sequence(avctx, seq);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_WARNING, "Failed to set decoder context.\n");
|
|
goto end;
|
|
}
|
|
|
|
end:
|
|
ff_cbs_fragment_reset(&s->current_obu);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int av1_frame_alloc(AVCodecContext *avctx, AV1Frame *f)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
AV1RawFrameHeader *header= s->raw_frame_header;
|
|
AVFrame *frame;
|
|
int ret;
|
|
|
|
ret = update_context_with_frame_header(avctx, header);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to update context with frame header\n");
|
|
return ret;
|
|
}
|
|
|
|
f->header_ref = av_buffer_ref(s->header_ref);
|
|
if (!f->header_ref)
|
|
return AVERROR(ENOMEM);
|
|
|
|
f->raw_frame_header = s->raw_frame_header;
|
|
|
|
if ((ret = ff_thread_get_buffer(avctx, &f->tf, AV_GET_BUFFER_FLAG_REF)) < 0)
|
|
goto fail;
|
|
|
|
frame = f->tf.f;
|
|
frame->key_frame = header->frame_type == AV1_FRAME_KEY;
|
|
|
|
switch (header->frame_type) {
|
|
case AV1_FRAME_KEY:
|
|
case AV1_FRAME_INTRA_ONLY:
|
|
frame->pict_type = AV_PICTURE_TYPE_I;
|
|
break;
|
|
case AV1_FRAME_INTER:
|
|
frame->pict_type = AV_PICTURE_TYPE_P;
|
|
break;
|
|
case AV1_FRAME_SWITCH:
|
|
frame->pict_type = AV_PICTURE_TYPE_SP;
|
|
break;
|
|
}
|
|
|
|
if (avctx->hwaccel) {
|
|
const AVHWAccel *hwaccel = avctx->hwaccel;
|
|
if (hwaccel->frame_priv_data_size) {
|
|
f->hwaccel_priv_buf =
|
|
av_buffer_allocz(hwaccel->frame_priv_data_size);
|
|
if (!f->hwaccel_priv_buf) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto fail;
|
|
}
|
|
f->hwaccel_picture_private = f->hwaccel_priv_buf->data;
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
fail:
|
|
av1_frame_unref(avctx, f);
|
|
return ret;
|
|
}
|
|
|
|
static int export_film_grain(AVCodecContext *avctx, AVFrame *frame)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
const AV1RawFilmGrainParams *film_grain = &s->cur_frame.film_grain;
|
|
AVFilmGrainParams *fgp;
|
|
AVFilmGrainAOMParams *aom;
|
|
|
|
if (!film_grain->apply_grain)
|
|
return 0;
|
|
|
|
fgp = av_film_grain_params_create_side_data(frame);
|
|
if (!fgp)
|
|
return AVERROR(ENOMEM);
|
|
|
|
fgp->type = AV_FILM_GRAIN_PARAMS_AV1;
|
|
fgp->seed = film_grain->grain_seed;
|
|
|
|
aom = &fgp->codec.aom;
|
|
aom->chroma_scaling_from_luma = film_grain->chroma_scaling_from_luma;
|
|
aom->scaling_shift = film_grain->grain_scaling_minus_8 + 8;
|
|
aom->ar_coeff_lag = film_grain->ar_coeff_lag;
|
|
aom->ar_coeff_shift = film_grain->ar_coeff_shift_minus_6 + 6;
|
|
aom->grain_scale_shift = film_grain->grain_scale_shift;
|
|
aom->overlap_flag = film_grain->overlap_flag;
|
|
aom->limit_output_range = film_grain->clip_to_restricted_range;
|
|
|
|
aom->num_y_points = film_grain->num_y_points;
|
|
for (int i = 0; i < film_grain->num_y_points; i++) {
|
|
aom->y_points[i][0] = film_grain->point_y_value[i];
|
|
aom->y_points[i][1] = film_grain->point_y_scaling[i];
|
|
}
|
|
aom->num_uv_points[0] = film_grain->num_cb_points;
|
|
for (int i = 0; i < film_grain->num_cb_points; i++) {
|
|
aom->uv_points[0][i][0] = film_grain->point_cb_value[i];
|
|
aom->uv_points[0][i][1] = film_grain->point_cb_scaling[i];
|
|
}
|
|
aom->num_uv_points[1] = film_grain->num_cr_points;
|
|
for (int i = 0; i < film_grain->num_cr_points; i++) {
|
|
aom->uv_points[1][i][0] = film_grain->point_cr_value[i];
|
|
aom->uv_points[1][i][1] = film_grain->point_cr_scaling[i];
|
|
}
|
|
|
|
for (int i = 0; i < 24; i++) {
|
|
aom->ar_coeffs_y[i] = film_grain->ar_coeffs_y_plus_128[i] - 128;
|
|
}
|
|
for (int i = 0; i < 25; i++) {
|
|
aom->ar_coeffs_uv[0][i] = film_grain->ar_coeffs_cb_plus_128[i] - 128;
|
|
aom->ar_coeffs_uv[1][i] = film_grain->ar_coeffs_cr_plus_128[i] - 128;
|
|
}
|
|
|
|
aom->uv_mult[0] = film_grain->cb_mult;
|
|
aom->uv_mult[1] = film_grain->cr_mult;
|
|
aom->uv_mult_luma[0] = film_grain->cb_luma_mult;
|
|
aom->uv_mult_luma[1] = film_grain->cr_luma_mult;
|
|
aom->uv_offset[0] = film_grain->cb_offset;
|
|
aom->uv_offset[1] = film_grain->cr_offset;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int set_output_frame(AVCodecContext *avctx, AVFrame *frame,
|
|
const AVPacket *pkt, int *got_frame)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
const AVFrame *srcframe = s->cur_frame.tf.f;
|
|
int ret;
|
|
|
|
// TODO: all layers
|
|
if (s->operating_point_idc &&
|
|
av_log2(s->operating_point_idc >> 8) > s->cur_frame.spatial_id)
|
|
return 0;
|
|
|
|
ret = av_frame_ref(frame, srcframe);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
if (avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) {
|
|
ret = export_film_grain(avctx, frame);
|
|
if (ret < 0) {
|
|
av_frame_unref(frame);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
frame->pts = pkt->pts;
|
|
frame->pkt_dts = pkt->dts;
|
|
frame->pkt_size = pkt->size;
|
|
|
|
*got_frame = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int update_reference_list(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
const AV1RawFrameHeader *header = s->raw_frame_header;
|
|
int ret;
|
|
|
|
for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
|
|
if (header->refresh_frame_flags & (1 << i)) {
|
|
if (s->ref[i].tf.f->buf[0])
|
|
av1_frame_unref(avctx, &s->ref[i]);
|
|
if ((ret = av1_frame_ref(avctx, &s->ref[i], &s->cur_frame)) < 0) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Failed to update frame %d in reference list\n", i);
|
|
return ret;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int get_current_frame(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
int ret;
|
|
|
|
if (s->cur_frame.tf.f->buf[0])
|
|
av1_frame_unref(avctx, &s->cur_frame);
|
|
|
|
ret = av1_frame_alloc(avctx, &s->cur_frame);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Failed to allocate space for current frame.\n");
|
|
return ret;
|
|
}
|
|
|
|
ret = init_tile_data(s);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to init tile data.\n");
|
|
return ret;
|
|
}
|
|
|
|
global_motion_params(s);
|
|
skip_mode_params(s);
|
|
coded_lossless_param(s);
|
|
load_grain_params(s);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int av1_decode_frame(AVCodecContext *avctx, void *frame,
|
|
int *got_frame, AVPacket *pkt)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
AV1RawTileGroup *raw_tile_group = NULL;
|
|
int ret;
|
|
|
|
ret = ff_cbs_read_packet(s->cbc, &s->current_obu, pkt);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to read packet.\n");
|
|
goto end;
|
|
}
|
|
av_log(avctx, AV_LOG_DEBUG, "Total obu for this frame:%d.\n",
|
|
s->current_obu.nb_units);
|
|
|
|
for (int i = 0; i < s->current_obu.nb_units; i++) {
|
|
CodedBitstreamUnit *unit = &s->current_obu.units[i];
|
|
AV1RawOBU *obu = unit->content;
|
|
const AV1RawOBUHeader *header;
|
|
|
|
if (!obu)
|
|
continue;
|
|
|
|
header = &obu->header;
|
|
av_log(avctx, AV_LOG_DEBUG, "Obu idx:%d, obu type:%d.\n", i, unit->type);
|
|
|
|
switch (unit->type) {
|
|
case AV1_OBU_SEQUENCE_HEADER:
|
|
av_buffer_unref(&s->seq_ref);
|
|
s->seq_ref = av_buffer_ref(unit->content_ref);
|
|
if (!s->seq_ref) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto end;
|
|
}
|
|
|
|
s->raw_seq = &obu->obu.sequence_header;
|
|
|
|
ret = set_context_with_sequence(avctx, s->raw_seq);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to set context.\n");
|
|
s->raw_seq = NULL;
|
|
goto end;
|
|
}
|
|
|
|
s->operating_point_idc = s->raw_seq->operating_point_idc[s->operating_point];
|
|
|
|
if (s->pix_fmt == AV_PIX_FMT_NONE) {
|
|
ret = get_pixel_format(avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"Failed to get pixel format.\n");
|
|
s->raw_seq = NULL;
|
|
goto end;
|
|
}
|
|
}
|
|
|
|
if (avctx->hwaccel && avctx->hwaccel->decode_params) {
|
|
ret = avctx->hwaccel->decode_params(avctx, unit->type, unit->data,
|
|
unit->data_size);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "HW accel decode params fail.\n");
|
|
s->raw_seq = NULL;
|
|
goto end;
|
|
}
|
|
}
|
|
break;
|
|
case AV1_OBU_REDUNDANT_FRAME_HEADER:
|
|
if (s->raw_frame_header)
|
|
break;
|
|
// fall-through
|
|
case AV1_OBU_FRAME:
|
|
case AV1_OBU_FRAME_HEADER:
|
|
if (!s->raw_seq) {
|
|
av_log(avctx, AV_LOG_ERROR, "Missing Sequence Header.\n");
|
|
ret = AVERROR_INVALIDDATA;
|
|
goto end;
|
|
}
|
|
|
|
av_buffer_unref(&s->header_ref);
|
|
s->header_ref = av_buffer_ref(unit->content_ref);
|
|
if (!s->header_ref) {
|
|
ret = AVERROR(ENOMEM);
|
|
goto end;
|
|
}
|
|
|
|
if (unit->type == AV1_OBU_FRAME)
|
|
s->raw_frame_header = &obu->obu.frame.header;
|
|
else
|
|
s->raw_frame_header = &obu->obu.frame_header;
|
|
|
|
if (s->raw_frame_header->show_existing_frame) {
|
|
if (s->cur_frame.tf.f->buf[0])
|
|
av1_frame_unref(avctx, &s->cur_frame);
|
|
|
|
ret = av1_frame_ref(avctx, &s->cur_frame,
|
|
&s->ref[s->raw_frame_header->frame_to_show_map_idx]);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to get reference frame.\n");
|
|
goto end;
|
|
}
|
|
|
|
ret = update_reference_list(avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to update reference list.\n");
|
|
goto end;
|
|
}
|
|
|
|
ret = set_output_frame(avctx, frame, pkt, got_frame);
|
|
if (ret < 0)
|
|
av_log(avctx, AV_LOG_ERROR, "Set output frame error.\n");
|
|
|
|
s->raw_frame_header = NULL;
|
|
|
|
goto end;
|
|
}
|
|
|
|
ret = get_current_frame(avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Get current frame error\n");
|
|
goto end;
|
|
}
|
|
|
|
s->cur_frame.spatial_id = header->spatial_id;
|
|
s->cur_frame.temporal_id = header->temporal_id;
|
|
|
|
if (avctx->hwaccel) {
|
|
ret = avctx->hwaccel->start_frame(avctx, unit->data,
|
|
unit->data_size);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "HW accel start frame fail.\n");
|
|
goto end;
|
|
}
|
|
}
|
|
if (unit->type != AV1_OBU_FRAME)
|
|
break;
|
|
// fall-through
|
|
case AV1_OBU_TILE_GROUP:
|
|
if (!s->raw_frame_header) {
|
|
av_log(avctx, AV_LOG_ERROR, "Missing Frame Header.\n");
|
|
ret = AVERROR_INVALIDDATA;
|
|
goto end;
|
|
}
|
|
|
|
if (unit->type == AV1_OBU_FRAME)
|
|
raw_tile_group = &obu->obu.frame.tile_group;
|
|
else
|
|
raw_tile_group = &obu->obu.tile_group;
|
|
|
|
ret = get_tiles_info(avctx, raw_tile_group);
|
|
if (ret < 0)
|
|
goto end;
|
|
|
|
if (avctx->hwaccel) {
|
|
ret = avctx->hwaccel->decode_slice(avctx,
|
|
raw_tile_group->tile_data.data,
|
|
raw_tile_group->tile_data.data_size);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR,
|
|
"HW accel decode slice fail.\n");
|
|
goto end;
|
|
}
|
|
}
|
|
break;
|
|
case AV1_OBU_TILE_LIST:
|
|
case AV1_OBU_TEMPORAL_DELIMITER:
|
|
case AV1_OBU_PADDING:
|
|
case AV1_OBU_METADATA:
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_DEBUG,
|
|
"Unknown obu type: %d (%"SIZE_SPECIFIER" bits).\n",
|
|
unit->type, unit->data_size);
|
|
}
|
|
|
|
if (raw_tile_group && (s->tile_num == raw_tile_group->tg_end + 1)) {
|
|
if (avctx->hwaccel) {
|
|
ret = avctx->hwaccel->end_frame(avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "HW accel end frame fail.\n");
|
|
goto end;
|
|
}
|
|
}
|
|
|
|
ret = update_reference_list(avctx);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to update reference list.\n");
|
|
goto end;
|
|
}
|
|
|
|
if (s->raw_frame_header->show_frame) {
|
|
ret = set_output_frame(avctx, frame, pkt, got_frame);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Set output frame error\n");
|
|
goto end;
|
|
}
|
|
}
|
|
raw_tile_group = NULL;
|
|
s->raw_frame_header = NULL;
|
|
}
|
|
}
|
|
|
|
end:
|
|
ff_cbs_fragment_reset(&s->current_obu);
|
|
if (ret < 0)
|
|
s->raw_frame_header = NULL;
|
|
return ret;
|
|
}
|
|
|
|
static void av1_decode_flush(AVCodecContext *avctx)
|
|
{
|
|
AV1DecContext *s = avctx->priv_data;
|
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++)
|
|
av1_frame_unref(avctx, &s->ref[i]);
|
|
|
|
av1_frame_unref(avctx, &s->cur_frame);
|
|
s->operating_point_idc = 0;
|
|
s->raw_frame_header = NULL;
|
|
s->raw_seq = NULL;
|
|
|
|
ff_cbs_flush(s->cbc);
|
|
}
|
|
|
|
#define OFFSET(x) offsetof(AV1DecContext, x)
|
|
#define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM
|
|
static const AVOption av1_options[] = {
|
|
{ "operating_point", "Select an operating point of the scalable bitstream",
|
|
OFFSET(operating_point), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, AV1_MAX_OPERATING_POINTS - 1, VD },
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVClass av1_class = {
|
|
.class_name = "AV1 decoder",
|
|
.item_name = av_default_item_name,
|
|
.option = av1_options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
const AVCodec ff_av1_decoder = {
|
|
.name = "av1",
|
|
.long_name = NULL_IF_CONFIG_SMALL("Alliance for Open Media AV1"),
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.id = AV_CODEC_ID_AV1,
|
|
.priv_data_size = sizeof(AV1DecContext),
|
|
.init = av1_decode_init,
|
|
.close = av1_decode_free,
|
|
.decode = av1_decode_frame,
|
|
.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_AVOID_PROBING,
|
|
.caps_internal = FF_CODEC_CAP_INIT_THREADSAFE |
|
|
FF_CODEC_CAP_INIT_CLEANUP |
|
|
FF_CODEC_CAP_SETS_PKT_DTS,
|
|
.flush = av1_decode_flush,
|
|
.profiles = NULL_IF_CONFIG_SMALL(ff_av1_profiles),
|
|
.priv_class = &av1_class,
|
|
.hw_configs = (const AVCodecHWConfigInternal *const []) {
|
|
#if CONFIG_AV1_DXVA2_HWACCEL
|
|
HWACCEL_DXVA2(av1),
|
|
#endif
|
|
#if CONFIG_AV1_D3D11VA_HWACCEL
|
|
HWACCEL_D3D11VA(av1),
|
|
#endif
|
|
#if CONFIG_AV1_D3D11VA2_HWACCEL
|
|
HWACCEL_D3D11VA2(av1),
|
|
#endif
|
|
#if CONFIG_AV1_NVDEC_HWACCEL
|
|
HWACCEL_NVDEC(av1),
|
|
#endif
|
|
#if CONFIG_AV1_VAAPI_HWACCEL
|
|
HWACCEL_VAAPI(av1),
|
|
#endif
|
|
NULL
|
|
},
|
|
};
|