1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-02 03:06:28 +02:00
FFmpeg/libavcodec/vaapi_av1.c
David Rosca d2d911eb9a lavc/vaapi_av1: Avoid sending the same slice buffer multiple times
When there are multiple tiles in one slice buffer, use multiple slice
params to avoid sending the same slice buffer multiple times and thus
increasing the bitstream size the driver will need to upload to hw.

Reviewed-by: Neal Gompa <ngompa13@gmail.com>
Signed-off-by: David Rosca <nowrep@gmail.com>
Signed-off-by: Haihao Xiang <haihao.xiang@intel.com>
2024-05-29 10:49:35 +08:00

459 lines
20 KiB
C

/*
* AV1 HW decode acceleration through VA API
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/frame.h"
#include "libavutil/mem.h"
#include "hwaccel_internal.h"
#include "vaapi_decode.h"
#include "internal.h"
#include "av1dec.h"
#include "thread.h"
typedef struct VAAPIAV1FrameRef {
AVFrame *frame;
int valid;
} VAAPIAV1FrameRef;
typedef struct VAAPIAV1DecContext {
VAAPIDecodeContext base;
/**
* For film grain case, VAAPI generate 2 output for each frame,
* current_frame will not apply film grain, and will be used for
* references for next frames. Maintain the reference list without
* applying film grain here. And current_display_picture will be
* used to apply film grain and push to downstream.
*/
VAAPIAV1FrameRef ref_tab[AV1_NUM_REF_FRAMES];
AVFrame *tmp_frame;
int nb_slice_params;
VASliceParameterBufferAV1 *slice_params;
} VAAPIAV1DecContext;
static VASurfaceID vaapi_av1_surface_id(AV1Frame *vf)
{
if (vf->f)
return ff_vaapi_get_surface_id(vf->f);
else
return VA_INVALID_SURFACE;
}
static int8_t vaapi_av1_get_bit_depth_idx(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
const AV1RawSequenceHeader *seq = s->raw_seq;
int8_t bit_depth = 8;
if (seq->seq_profile == 2 && seq->color_config.high_bitdepth)
bit_depth = seq->color_config.twelve_bit ? 12 : 10;
else if (seq->seq_profile <= 2)
bit_depth = seq->color_config.high_bitdepth ? 10 : 8;
else {
av_log(avctx, AV_LOG_ERROR,
"Couldn't get bit depth from profile:%d.\n", seq->seq_profile);
return -1;
}
return bit_depth == 8 ? 0 : bit_depth == 10 ? 1 : 2;
}
static int vaapi_av1_decode_init(AVCodecContext *avctx)
{
VAAPIAV1DecContext *ctx = avctx->internal->hwaccel_priv_data;
ctx->tmp_frame = av_frame_alloc();
if (!ctx->tmp_frame)
return AVERROR(ENOMEM);
for (int i = 0; i < FF_ARRAY_ELEMS(ctx->ref_tab); i++) {
ctx->ref_tab[i].frame = av_frame_alloc();
if (!ctx->ref_tab[i].frame)
return AVERROR(ENOMEM);
ctx->ref_tab[i].valid = 0;
}
return ff_vaapi_decode_init(avctx);
}
static int vaapi_av1_decode_uninit(AVCodecContext *avctx)
{
VAAPIAV1DecContext *ctx = avctx->internal->hwaccel_priv_data;
av_frame_free(&ctx->tmp_frame);
for (int i = 0; i < FF_ARRAY_ELEMS(ctx->ref_tab); i++)
av_frame_free(&ctx->ref_tab[i].frame);
av_freep(&ctx->slice_params);
return ff_vaapi_decode_uninit(avctx);
}
static int vaapi_av1_start_frame(AVCodecContext *avctx,
av_unused const uint8_t *buffer,
av_unused uint32_t size)
{
AV1DecContext *s = avctx->priv_data;
const AV1RawSequenceHeader *seq = s->raw_seq;
const AV1RawFrameHeader *frame_header = s->raw_frame_header;
const AV1RawFilmGrainParams *film_grain = &s->cur_frame.film_grain;
VAAPIDecodePicture *pic = s->cur_frame.hwaccel_picture_private;
VAAPIAV1DecContext *ctx = avctx->internal->hwaccel_priv_data;
VADecPictureParameterBufferAV1 pic_param;
int8_t bit_depth_idx;
int err = 0;
int apply_grain = !(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) && film_grain->apply_grain;
uint8_t remap_lr_type[4] = {AV1_RESTORE_NONE, AV1_RESTORE_SWITCHABLE, AV1_RESTORE_WIENER, AV1_RESTORE_SGRPROJ};
uint8_t segmentation_feature_signed[AV1_SEG_LVL_MAX] = {1, 1, 1, 1, 1, 0, 0, 0};
uint8_t segmentation_feature_max[AV1_SEG_LVL_MAX] = {255, AV1_MAX_LOOP_FILTER,
AV1_MAX_LOOP_FILTER, AV1_MAX_LOOP_FILTER, AV1_MAX_LOOP_FILTER, 7 , 0 , 0 };
bit_depth_idx = vaapi_av1_get_bit_depth_idx(avctx);
if (bit_depth_idx < 0)
goto fail;
if (apply_grain) {
if (ctx->tmp_frame->buf[0])
av_frame_unref(ctx->tmp_frame);
err = ff_thread_get_buffer(avctx, ctx->tmp_frame, AV_GET_BUFFER_FLAG_REF);
if (err < 0)
goto fail;
pic->output_surface = ff_vaapi_get_surface_id(ctx->tmp_frame);
} else {
pic->output_surface = ff_vaapi_get_surface_id(s->cur_frame.f);
}
memset(&pic_param, 0, sizeof(VADecPictureParameterBufferAV1));
pic_param = (VADecPictureParameterBufferAV1) {
.profile = seq->seq_profile,
.order_hint_bits_minus_1 = seq->order_hint_bits_minus_1,
.bit_depth_idx = bit_depth_idx,
.matrix_coefficients = seq->color_config.matrix_coefficients,
.current_frame = pic->output_surface,
.current_display_picture = ff_vaapi_get_surface_id(s->cur_frame.f),
.frame_width_minus1 = frame_header->frame_width_minus_1,
.frame_height_minus1 = frame_header->frame_height_minus_1,
.primary_ref_frame = frame_header->primary_ref_frame,
.order_hint = frame_header->order_hint,
.tile_cols = frame_header->tile_cols,
.tile_rows = frame_header->tile_rows,
.context_update_tile_id = frame_header->context_update_tile_id,
.superres_scale_denominator = frame_header->use_superres ?
frame_header->coded_denom + AV1_SUPERRES_DENOM_MIN :
AV1_SUPERRES_NUM,
.interp_filter = frame_header->interpolation_filter,
.filter_level[0] = frame_header->loop_filter_level[0],
.filter_level[1] = frame_header->loop_filter_level[1],
.filter_level_u = frame_header->loop_filter_level[2],
.filter_level_v = frame_header->loop_filter_level[3],
.base_qindex = frame_header->base_q_idx,
.y_dc_delta_q = frame_header->delta_q_y_dc,
.u_dc_delta_q = frame_header->delta_q_u_dc,
.u_ac_delta_q = frame_header->delta_q_u_ac,
.v_dc_delta_q = frame_header->delta_q_v_dc,
.v_ac_delta_q = frame_header->delta_q_v_ac,
.cdef_damping_minus_3 = frame_header->cdef_damping_minus_3,
.cdef_bits = frame_header->cdef_bits,
.seq_info_fields.fields = {
.still_picture = seq->still_picture,
.use_128x128_superblock = seq->use_128x128_superblock,
.enable_filter_intra = seq->enable_filter_intra,
.enable_intra_edge_filter = seq->enable_intra_edge_filter,
.enable_interintra_compound = seq->enable_interintra_compound,
.enable_masked_compound = seq->enable_masked_compound,
.enable_dual_filter = seq->enable_dual_filter,
.enable_order_hint = seq->enable_order_hint,
.enable_jnt_comp = seq->enable_jnt_comp,
.enable_cdef = seq->enable_cdef,
.mono_chrome = seq->color_config.mono_chrome,
.color_range = seq->color_config.color_range,
.subsampling_x = seq->color_config.subsampling_x,
.subsampling_y = seq->color_config.subsampling_y,
.chroma_sample_position = seq->color_config.chroma_sample_position,
.film_grain_params_present = seq->film_grain_params_present &&
!(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN),
},
.seg_info.segment_info_fields.bits = {
.enabled = frame_header->segmentation_enabled,
.update_map = frame_header->segmentation_update_map,
.temporal_update = frame_header->segmentation_temporal_update,
.update_data = frame_header->segmentation_update_data,
},
.film_grain_info = {
.film_grain_info_fields.bits = {
.apply_grain = apply_grain,
.chroma_scaling_from_luma = film_grain->chroma_scaling_from_luma,
.grain_scaling_minus_8 = film_grain->grain_scaling_minus_8,
.ar_coeff_lag = film_grain->ar_coeff_lag,
.ar_coeff_shift_minus_6 = film_grain->ar_coeff_shift_minus_6,
.grain_scale_shift = film_grain->grain_scale_shift,
.overlap_flag = film_grain->overlap_flag,
.clip_to_restricted_range = film_grain->clip_to_restricted_range,
},
.grain_seed = film_grain->grain_seed,
.num_y_points = film_grain->num_y_points,
.num_cb_points = film_grain->num_cb_points,
.num_cr_points = film_grain->num_cr_points,
.cb_mult = film_grain->cb_mult,
.cb_luma_mult = film_grain->cb_luma_mult,
.cb_offset = film_grain->cb_offset,
.cr_mult = film_grain->cr_mult,
.cr_luma_mult = film_grain->cr_luma_mult,
.cr_offset = film_grain->cr_offset,
},
.pic_info_fields.bits = {
.frame_type = frame_header->frame_type,
.show_frame = frame_header->show_frame,
.showable_frame = frame_header->showable_frame,
.error_resilient_mode = frame_header->error_resilient_mode,
.disable_cdf_update = frame_header->disable_cdf_update,
.allow_screen_content_tools = frame_header->allow_screen_content_tools,
.force_integer_mv = s->cur_frame.force_integer_mv,
.allow_intrabc = frame_header->allow_intrabc,
.use_superres = frame_header->use_superres,
.allow_high_precision_mv = frame_header->allow_high_precision_mv,
.is_motion_mode_switchable = frame_header->is_motion_mode_switchable,
.use_ref_frame_mvs = frame_header->use_ref_frame_mvs,
.disable_frame_end_update_cdf = frame_header->disable_frame_end_update_cdf,
.uniform_tile_spacing_flag = frame_header->uniform_tile_spacing_flag,
.allow_warped_motion = frame_header->allow_warped_motion,
},
.loop_filter_info_fields.bits = {
.sharpness_level = frame_header->loop_filter_sharpness,
.mode_ref_delta_enabled = frame_header->loop_filter_delta_enabled,
.mode_ref_delta_update = frame_header->loop_filter_delta_update,
},
.mode_control_fields.bits = {
.delta_q_present_flag = frame_header->delta_q_present,
.log2_delta_q_res = frame_header->delta_q_res,
.delta_lf_present_flag = frame_header->delta_lf_present,
.log2_delta_lf_res = frame_header->delta_lf_res,
.delta_lf_multi = frame_header->delta_lf_multi,
.tx_mode = frame_header->tx_mode,
.reference_select = frame_header->reference_select,
.reduced_tx_set_used = frame_header->reduced_tx_set,
.skip_mode_present = frame_header->skip_mode_present,
},
.loop_restoration_fields.bits = {
.yframe_restoration_type = remap_lr_type[frame_header->lr_type[0]],
.cbframe_restoration_type = remap_lr_type[frame_header->lr_type[1]],
.crframe_restoration_type = remap_lr_type[frame_header->lr_type[2]],
.lr_unit_shift = frame_header->lr_unit_shift,
.lr_uv_shift = frame_header->lr_uv_shift,
},
.qmatrix_fields.bits = {
.using_qmatrix = frame_header->using_qmatrix,
.qm_y = frame_header->qm_y,
.qm_u = frame_header->qm_u,
.qm_v = frame_header->qm_v,
}
};
for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
if (pic_param.pic_info_fields.bits.frame_type == AV1_FRAME_KEY && frame_header->show_frame)
pic_param.ref_frame_map[i] = VA_INVALID_ID;
else
pic_param.ref_frame_map[i] = ctx->ref_tab[i].valid ?
ff_vaapi_get_surface_id(ctx->ref_tab[i].frame) :
vaapi_av1_surface_id(&s->ref[i]);
}
for (int i = 0; i < AV1_REFS_PER_FRAME; i++) {
pic_param.ref_frame_idx[i] = frame_header->ref_frame_idx[i];
}
for (int i = 0; i < AV1_TOTAL_REFS_PER_FRAME; i++) {
pic_param.ref_deltas[i] = frame_header->loop_filter_ref_deltas[i];
}
for (int i = 0; i < 2; i++) {
pic_param.mode_deltas[i] = frame_header->loop_filter_mode_deltas[i];
}
for (int i = 0; i < (1 << frame_header->cdef_bits); i++) {
pic_param.cdef_y_strengths[i] =
(frame_header->cdef_y_pri_strength[i] << 2) +
frame_header->cdef_y_sec_strength[i];
pic_param.cdef_uv_strengths[i] =
(frame_header->cdef_uv_pri_strength[i] << 2) +
frame_header->cdef_uv_sec_strength[i];
}
for (int i = 0; i < frame_header->tile_cols; i++) {
pic_param.width_in_sbs_minus_1[i] =
frame_header->width_in_sbs_minus_1[i];
}
for (int i = 0; i < frame_header->tile_rows; i++) {
pic_param.height_in_sbs_minus_1[i] =
frame_header->height_in_sbs_minus_1[i];
}
for (int i = AV1_REF_FRAME_LAST; i <= AV1_REF_FRAME_ALTREF; i++) {
pic_param.wm[i - 1].invalid = s->cur_frame.gm_invalid[i];
pic_param.wm[i - 1].wmtype = s->cur_frame.gm_type[i];
for (int j = 0; j < 6; j++)
pic_param.wm[i - 1].wmmat[j] = s->cur_frame.gm_params[i][j];
}
for (int i = 0; i < AV1_MAX_SEGMENTS; i++) {
for (int j = 0; j < AV1_SEG_LVL_MAX; j++) {
pic_param.seg_info.feature_mask[i] |= (frame_header->feature_enabled[i][j] << j);
if (segmentation_feature_signed[j])
pic_param.seg_info.feature_data[i][j] = av_clip(frame_header->feature_value[i][j],
-segmentation_feature_max[j], segmentation_feature_max[j]);
else
pic_param.seg_info.feature_data[i][j] = av_clip(frame_header->feature_value[i][j],
0, segmentation_feature_max[j]);
}
}
if (apply_grain) {
for (int i = 0; i < film_grain->num_y_points; i++) {
pic_param.film_grain_info.point_y_value[i] =
film_grain->point_y_value[i];
pic_param.film_grain_info.point_y_scaling[i] =
film_grain->point_y_scaling[i];
}
for (int i = 0; i < film_grain->num_cb_points; i++) {
pic_param.film_grain_info.point_cb_value[i] =
film_grain->point_cb_value[i];
pic_param.film_grain_info.point_cb_scaling[i] =
film_grain->point_cb_scaling[i];
}
for (int i = 0; i < film_grain->num_cr_points; i++) {
pic_param.film_grain_info.point_cr_value[i] =
film_grain->point_cr_value[i];
pic_param.film_grain_info.point_cr_scaling[i] =
film_grain->point_cr_scaling[i];
}
for (int i = 0; i < 24; i++) {
pic_param.film_grain_info.ar_coeffs_y[i] =
film_grain->ar_coeffs_y_plus_128[i] - 128;
}
for (int i = 0; i < 25; i++) {
pic_param.film_grain_info.ar_coeffs_cb[i] =
film_grain->ar_coeffs_cb_plus_128[i] - 128;
pic_param.film_grain_info.ar_coeffs_cr[i] =
film_grain->ar_coeffs_cr_plus_128[i] - 128;
}
}
err = ff_vaapi_decode_make_param_buffer(avctx, pic,
VAPictureParameterBufferType,
&pic_param, sizeof(pic_param));
if (err < 0)
goto fail;
return 0;
fail:
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
static int vaapi_av1_end_frame(AVCodecContext *avctx)
{
const AV1DecContext *s = avctx->priv_data;
const AV1RawFrameHeader *header = s->raw_frame_header;
const AV1RawFilmGrainParams *film_grain = &s->cur_frame.film_grain;
VAAPIDecodePicture *pic = s->cur_frame.hwaccel_picture_private;
VAAPIAV1DecContext *ctx = avctx->internal->hwaccel_priv_data;
int apply_grain = !(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) && film_grain->apply_grain;
int ret;
ret = ff_vaapi_decode_issue(avctx, pic);
if (ret < 0)
return ret;
for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
if (header->refresh_frame_flags & (1 << i)) {
if (ctx->ref_tab[i].frame->buf[0])
av_frame_unref(ctx->ref_tab[i].frame);
if (apply_grain) {
ret = av_frame_ref(ctx->ref_tab[i].frame, ctx->tmp_frame);
if (ret < 0)
return ret;
ctx->ref_tab[i].valid = 1;
} else {
ctx->ref_tab[i].valid = 0;
}
}
}
return 0;
}
static int vaapi_av1_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer,
uint32_t size)
{
const AV1DecContext *s = avctx->priv_data;
VAAPIDecodePicture *pic = s->cur_frame.hwaccel_picture_private;
VAAPIAV1DecContext *ctx = avctx->internal->hwaccel_priv_data;
int err, nb_params;
nb_params = s->tg_end - s->tg_start + 1;
if (ctx->nb_slice_params < nb_params) {
ctx->slice_params = av_realloc_array(ctx->slice_params,
nb_params,
sizeof(*ctx->slice_params));
if (!ctx->slice_params) {
ctx->nb_slice_params = 0;
err = AVERROR(ENOMEM);
goto fail;
}
ctx->nb_slice_params = nb_params;
}
for (int i = s->tg_start; i <= s->tg_end; i++) {
ctx->slice_params[i - s->tg_start] = (VASliceParameterBufferAV1) {
.slice_data_size = s->tile_group_info[i].tile_size,
.slice_data_offset = s->tile_group_info[i].tile_offset,
.slice_data_flag = VA_SLICE_DATA_FLAG_ALL,
.tile_row = s->tile_group_info[i].tile_row,
.tile_column = s->tile_group_info[i].tile_column,
.tg_start = s->tg_start,
.tg_end = s->tg_end,
};
}
err = ff_vaapi_decode_make_slice_buffer(avctx, pic, ctx->slice_params, nb_params,
sizeof(VASliceParameterBufferAV1),
buffer,
size);
if (err)
goto fail;
return 0;
fail:
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
const FFHWAccel ff_av1_vaapi_hwaccel = {
.p.name = "av1_vaapi",
.p.type = AVMEDIA_TYPE_VIDEO,
.p.id = AV_CODEC_ID_AV1,
.p.pix_fmt = AV_PIX_FMT_VAAPI,
.start_frame = vaapi_av1_start_frame,
.end_frame = vaapi_av1_end_frame,
.decode_slice = vaapi_av1_decode_slice,
.frame_priv_data_size = sizeof(VAAPIDecodePicture),
.init = vaapi_av1_decode_init,
.uninit = vaapi_av1_decode_uninit,
.frame_params = ff_vaapi_common_frame_params,
.priv_data_size = sizeof(VAAPIAV1DecContext),
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
};