mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
5aabb2596f
This structure is no longer declared in a public header, so using an FF-prefix is more appropriate. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
1326 lines
48 KiB
C
1326 lines
48 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include <string.h>
|
|
|
|
#include <va/va.h>
|
|
#include <va/va_enc_hevc.h>
|
|
|
|
#include "libavutil/avassert.h"
|
|
#include "libavutil/common.h"
|
|
#include "libavutil/pixdesc.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/mastering_display_metadata.h"
|
|
|
|
#include "avcodec.h"
|
|
#include "cbs.h"
|
|
#include "cbs_h265.h"
|
|
#include "codec_internal.h"
|
|
#include "h265_profile_level.h"
|
|
#include "hevc.h"
|
|
#include "hevc_sei.h"
|
|
#include "put_bits.h"
|
|
#include "vaapi_encode.h"
|
|
|
|
enum {
|
|
SEI_MASTERING_DISPLAY = 0x08,
|
|
SEI_CONTENT_LIGHT_LEVEL = 0x10,
|
|
};
|
|
|
|
typedef struct VAAPIEncodeH265Picture {
|
|
int pic_order_cnt;
|
|
|
|
int64_t last_idr_frame;
|
|
|
|
int slice_nal_unit;
|
|
int slice_type;
|
|
int pic_type;
|
|
} VAAPIEncodeH265Picture;
|
|
|
|
typedef struct VAAPIEncodeH265Context {
|
|
VAAPIEncodeContext common;
|
|
|
|
// User options.
|
|
int qp;
|
|
int aud;
|
|
int profile;
|
|
int tier;
|
|
int level;
|
|
int sei;
|
|
|
|
// Derived settings.
|
|
int fixed_qp_idr;
|
|
int fixed_qp_p;
|
|
int fixed_qp_b;
|
|
|
|
// Writer structures.
|
|
H265RawAUD raw_aud;
|
|
H265RawVPS raw_vps;
|
|
H265RawSPS raw_sps;
|
|
H265RawPPS raw_pps;
|
|
H265RawSlice raw_slice;
|
|
|
|
SEIRawMasteringDisplayColourVolume sei_mastering_display;
|
|
SEIRawContentLightLevelInfo sei_content_light_level;
|
|
|
|
CodedBitstreamContext *cbc;
|
|
CodedBitstreamFragment current_access_unit;
|
|
int aud_needed;
|
|
int sei_needed;
|
|
} VAAPIEncodeH265Context;
|
|
|
|
|
|
static int vaapi_encode_h265_write_access_unit(AVCodecContext *avctx,
|
|
char *data, size_t *data_len,
|
|
CodedBitstreamFragment *au)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
int err;
|
|
|
|
err = ff_cbs_write_fragment_data(priv->cbc, au);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to write packed header.\n");
|
|
return err;
|
|
}
|
|
|
|
if (*data_len < 8 * au->data_size - au->data_bit_padding) {
|
|
av_log(avctx, AV_LOG_ERROR, "Access unit too large: "
|
|
"%zu < %zu.\n", *data_len,
|
|
8 * au->data_size - au->data_bit_padding);
|
|
return AVERROR(ENOSPC);
|
|
}
|
|
|
|
memcpy(data, au->data, au->data_size);
|
|
*data_len = 8 * au->data_size - au->data_bit_padding;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vaapi_encode_h265_add_nal(AVCodecContext *avctx,
|
|
CodedBitstreamFragment *au,
|
|
void *nal_unit)
|
|
{
|
|
H265RawNALUnitHeader *header = nal_unit;
|
|
int err;
|
|
|
|
err = ff_cbs_insert_unit_content(au, -1,
|
|
header->nal_unit_type, nal_unit, NULL);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to add NAL unit: "
|
|
"type = %d.\n", header->nal_unit_type);
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vaapi_encode_h265_write_sequence_header(AVCodecContext *avctx,
|
|
char *data, size_t *data_len)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
CodedBitstreamFragment *au = &priv->current_access_unit;
|
|
int err;
|
|
|
|
if (priv->aud_needed) {
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_aud);
|
|
if (err < 0)
|
|
goto fail;
|
|
priv->aud_needed = 0;
|
|
}
|
|
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_vps);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_sps);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_pps);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au);
|
|
fail:
|
|
ff_cbs_fragment_reset(au);
|
|
return err;
|
|
}
|
|
|
|
static int vaapi_encode_h265_write_slice_header(AVCodecContext *avctx,
|
|
VAAPIEncodePicture *pic,
|
|
VAAPIEncodeSlice *slice,
|
|
char *data, size_t *data_len)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
CodedBitstreamFragment *au = &priv->current_access_unit;
|
|
int err;
|
|
|
|
if (priv->aud_needed) {
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_aud);
|
|
if (err < 0)
|
|
goto fail;
|
|
priv->aud_needed = 0;
|
|
}
|
|
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_slice);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au);
|
|
fail:
|
|
ff_cbs_fragment_reset(au);
|
|
return err;
|
|
}
|
|
|
|
static int vaapi_encode_h265_write_extra_header(AVCodecContext *avctx,
|
|
VAAPIEncodePicture *pic,
|
|
int index, int *type,
|
|
char *data, size_t *data_len)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
CodedBitstreamFragment *au = &priv->current_access_unit;
|
|
int err;
|
|
|
|
if (priv->sei_needed) {
|
|
if (priv->aud_needed) {
|
|
err = vaapi_encode_h265_add_nal(avctx, au, &priv->aud);
|
|
if (err < 0)
|
|
goto fail;
|
|
priv->aud_needed = 0;
|
|
}
|
|
|
|
if (priv->sei_needed & SEI_MASTERING_DISPLAY) {
|
|
err = ff_cbs_sei_add_message(priv->cbc, au, 1,
|
|
SEI_TYPE_MASTERING_DISPLAY_COLOUR_VOLUME,
|
|
&priv->sei_mastering_display, NULL);
|
|
if (err < 0)
|
|
goto fail;
|
|
}
|
|
|
|
if (priv->sei_needed & SEI_CONTENT_LIGHT_LEVEL) {
|
|
err = ff_cbs_sei_add_message(priv->cbc, au, 1,
|
|
SEI_TYPE_CONTENT_LIGHT_LEVEL_INFO,
|
|
&priv->sei_content_light_level, NULL);
|
|
if (err < 0)
|
|
goto fail;
|
|
}
|
|
|
|
priv->sei_needed = 0;
|
|
|
|
err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
ff_cbs_fragment_reset(au);
|
|
|
|
*type = VAEncPackedHeaderRawData;
|
|
return 0;
|
|
} else {
|
|
return AVERROR_EOF;
|
|
}
|
|
|
|
fail:
|
|
ff_cbs_fragment_reset(au);
|
|
return err;
|
|
}
|
|
|
|
static int vaapi_encode_h265_init_sequence_params(AVCodecContext *avctx)
|
|
{
|
|
VAAPIEncodeContext *ctx = avctx->priv_data;
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
H265RawVPS *vps = &priv->raw_vps;
|
|
H265RawSPS *sps = &priv->raw_sps;
|
|
H265RawPPS *pps = &priv->raw_pps;
|
|
H265RawProfileTierLevel *ptl = &vps->profile_tier_level;
|
|
H265RawVUI *vui = &sps->vui;
|
|
VAEncSequenceParameterBufferHEVC *vseq = ctx->codec_sequence_params;
|
|
VAEncPictureParameterBufferHEVC *vpic = ctx->codec_picture_params;
|
|
const AVPixFmtDescriptor *desc;
|
|
int chroma_format, bit_depth;
|
|
int i;
|
|
|
|
memset(vps, 0, sizeof(*vps));
|
|
memset(sps, 0, sizeof(*sps));
|
|
memset(pps, 0, sizeof(*pps));
|
|
|
|
|
|
desc = av_pix_fmt_desc_get(priv->common.input_frames->sw_format);
|
|
av_assert0(desc);
|
|
if (desc->nb_components == 1) {
|
|
chroma_format = 0;
|
|
} else {
|
|
if (desc->log2_chroma_w == 1 && desc->log2_chroma_h == 1) {
|
|
chroma_format = 1;
|
|
} else if (desc->log2_chroma_w == 1 && desc->log2_chroma_h == 0) {
|
|
chroma_format = 2;
|
|
} else if (desc->log2_chroma_w == 0 && desc->log2_chroma_h == 0) {
|
|
chroma_format = 3;
|
|
} else {
|
|
av_log(avctx, AV_LOG_ERROR, "Chroma format of input pixel format "
|
|
"%s is not supported.\n", desc->name);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
}
|
|
bit_depth = desc->comp[0].depth;
|
|
|
|
|
|
// VPS
|
|
|
|
vps->nal_unit_header = (H265RawNALUnitHeader) {
|
|
.nal_unit_type = HEVC_NAL_VPS,
|
|
.nuh_layer_id = 0,
|
|
.nuh_temporal_id_plus1 = 1,
|
|
};
|
|
|
|
vps->vps_video_parameter_set_id = 0;
|
|
|
|
vps->vps_base_layer_internal_flag = 1;
|
|
vps->vps_base_layer_available_flag = 1;
|
|
vps->vps_max_layers_minus1 = 0;
|
|
vps->vps_max_sub_layers_minus1 = 0;
|
|
vps->vps_temporal_id_nesting_flag = 1;
|
|
|
|
ptl->general_profile_space = 0;
|
|
ptl->general_profile_idc = avctx->profile;
|
|
ptl->general_tier_flag = priv->tier;
|
|
|
|
if (chroma_format == 1) {
|
|
ptl->general_profile_compatibility_flag[1] = bit_depth == 8;
|
|
ptl->general_profile_compatibility_flag[2] = bit_depth <= 10;
|
|
}
|
|
ptl->general_profile_compatibility_flag[4] = 1;
|
|
|
|
ptl->general_progressive_source_flag = 1;
|
|
ptl->general_interlaced_source_flag = 0;
|
|
ptl->general_non_packed_constraint_flag = 1;
|
|
ptl->general_frame_only_constraint_flag = 1;
|
|
|
|
ptl->general_max_12bit_constraint_flag = bit_depth <= 12;
|
|
ptl->general_max_10bit_constraint_flag = bit_depth <= 10;
|
|
ptl->general_max_8bit_constraint_flag = bit_depth == 8;
|
|
|
|
ptl->general_max_422chroma_constraint_flag = chroma_format <= 2;
|
|
ptl->general_max_420chroma_constraint_flag = chroma_format <= 1;
|
|
ptl->general_max_monochrome_constraint_flag = chroma_format == 0;
|
|
|
|
ptl->general_intra_constraint_flag = ctx->gop_size == 1;
|
|
|
|
ptl->general_lower_bit_rate_constraint_flag = 1;
|
|
|
|
if (avctx->level != FF_LEVEL_UNKNOWN) {
|
|
ptl->general_level_idc = avctx->level;
|
|
} else {
|
|
const H265LevelDescriptor *level;
|
|
|
|
level = ff_h265_guess_level(ptl, avctx->bit_rate,
|
|
ctx->surface_width, ctx->surface_height,
|
|
ctx->nb_slices, ctx->tile_rows, ctx->tile_cols,
|
|
(ctx->b_per_p > 0) + 1);
|
|
if (level) {
|
|
av_log(avctx, AV_LOG_VERBOSE, "Using level %s.\n", level->name);
|
|
ptl->general_level_idc = level->level_idc;
|
|
} else {
|
|
av_log(avctx, AV_LOG_VERBOSE, "Stream will not conform to "
|
|
"any normal level; using level 8.5.\n");
|
|
ptl->general_level_idc = 255;
|
|
// The tier flag must be set in level 8.5.
|
|
ptl->general_tier_flag = 1;
|
|
}
|
|
}
|
|
|
|
vps->vps_sub_layer_ordering_info_present_flag = 0;
|
|
vps->vps_max_dec_pic_buffering_minus1[0] = ctx->max_b_depth + 1;
|
|
vps->vps_max_num_reorder_pics[0] = ctx->max_b_depth;
|
|
vps->vps_max_latency_increase_plus1[0] = 0;
|
|
|
|
vps->vps_max_layer_id = 0;
|
|
vps->vps_num_layer_sets_minus1 = 0;
|
|
vps->layer_id_included_flag[0][0] = 1;
|
|
|
|
vps->vps_timing_info_present_flag = 1;
|
|
if (avctx->framerate.num > 0 && avctx->framerate.den > 0) {
|
|
vps->vps_num_units_in_tick = avctx->framerate.den;
|
|
vps->vps_time_scale = avctx->framerate.num;
|
|
vps->vps_poc_proportional_to_timing_flag = 1;
|
|
vps->vps_num_ticks_poc_diff_one_minus1 = 0;
|
|
} else {
|
|
vps->vps_num_units_in_tick = avctx->time_base.num;
|
|
vps->vps_time_scale = avctx->time_base.den;
|
|
vps->vps_poc_proportional_to_timing_flag = 0;
|
|
}
|
|
vps->vps_num_hrd_parameters = 0;
|
|
|
|
|
|
// SPS
|
|
|
|
sps->nal_unit_header = (H265RawNALUnitHeader) {
|
|
.nal_unit_type = HEVC_NAL_SPS,
|
|
.nuh_layer_id = 0,
|
|
.nuh_temporal_id_plus1 = 1,
|
|
};
|
|
|
|
sps->sps_video_parameter_set_id = vps->vps_video_parameter_set_id;
|
|
|
|
sps->sps_max_sub_layers_minus1 = vps->vps_max_sub_layers_minus1;
|
|
sps->sps_temporal_id_nesting_flag = vps->vps_temporal_id_nesting_flag;
|
|
|
|
sps->profile_tier_level = vps->profile_tier_level;
|
|
|
|
sps->sps_seq_parameter_set_id = 0;
|
|
|
|
sps->chroma_format_idc = chroma_format;
|
|
sps->separate_colour_plane_flag = 0;
|
|
|
|
sps->pic_width_in_luma_samples = ctx->surface_width;
|
|
sps->pic_height_in_luma_samples = ctx->surface_height;
|
|
|
|
if (avctx->width != ctx->surface_width ||
|
|
avctx->height != ctx->surface_height) {
|
|
sps->conformance_window_flag = 1;
|
|
sps->conf_win_left_offset = 0;
|
|
sps->conf_win_right_offset =
|
|
(ctx->surface_width - avctx->width) >> desc->log2_chroma_w;
|
|
sps->conf_win_top_offset = 0;
|
|
sps->conf_win_bottom_offset =
|
|
(ctx->surface_height - avctx->height) >> desc->log2_chroma_h;
|
|
} else {
|
|
sps->conformance_window_flag = 0;
|
|
}
|
|
|
|
sps->bit_depth_luma_minus8 = bit_depth - 8;
|
|
sps->bit_depth_chroma_minus8 = bit_depth - 8;
|
|
|
|
sps->log2_max_pic_order_cnt_lsb_minus4 = 8;
|
|
|
|
sps->sps_sub_layer_ordering_info_present_flag =
|
|
vps->vps_sub_layer_ordering_info_present_flag;
|
|
for (i = 0; i <= sps->sps_max_sub_layers_minus1; i++) {
|
|
sps->sps_max_dec_pic_buffering_minus1[i] =
|
|
vps->vps_max_dec_pic_buffering_minus1[i];
|
|
sps->sps_max_num_reorder_pics[i] =
|
|
vps->vps_max_num_reorder_pics[i];
|
|
sps->sps_max_latency_increase_plus1[i] =
|
|
vps->vps_max_latency_increase_plus1[i];
|
|
}
|
|
|
|
// These have to come from the capabilities of the encoder. We have no
|
|
// way to query them, so just hardcode parameters which work on the Intel
|
|
// driver.
|
|
// CTB size from 8x8 to 32x32.
|
|
sps->log2_min_luma_coding_block_size_minus3 = 0;
|
|
sps->log2_diff_max_min_luma_coding_block_size = 2;
|
|
// Transform size from 4x4 to 32x32.
|
|
sps->log2_min_luma_transform_block_size_minus2 = 0;
|
|
sps->log2_diff_max_min_luma_transform_block_size = 3;
|
|
// Full transform hierarchy allowed (2-5).
|
|
sps->max_transform_hierarchy_depth_inter = 3;
|
|
sps->max_transform_hierarchy_depth_intra = 3;
|
|
// AMP works.
|
|
sps->amp_enabled_flag = 1;
|
|
// SAO and temporal MVP do not work.
|
|
sps->sample_adaptive_offset_enabled_flag = 0;
|
|
sps->sps_temporal_mvp_enabled_flag = 0;
|
|
|
|
sps->pcm_enabled_flag = 0;
|
|
|
|
// STRPSs should ideally be here rather than defined individually in
|
|
// each slice, but the structure isn't completely fixed so for now
|
|
// don't bother.
|
|
sps->num_short_term_ref_pic_sets = 0;
|
|
sps->long_term_ref_pics_present_flag = 0;
|
|
|
|
sps->vui_parameters_present_flag = 1;
|
|
|
|
if (avctx->sample_aspect_ratio.num != 0 &&
|
|
avctx->sample_aspect_ratio.den != 0) {
|
|
static const AVRational sar_idc[] = {
|
|
{ 0, 0 },
|
|
{ 1, 1 }, { 12, 11 }, { 10, 11 }, { 16, 11 },
|
|
{ 40, 33 }, { 24, 11 }, { 20, 11 }, { 32, 11 },
|
|
{ 80, 33 }, { 18, 11 }, { 15, 11 }, { 64, 33 },
|
|
{ 160, 99 }, { 4, 3 }, { 3, 2 }, { 2, 1 },
|
|
};
|
|
int num, den, i;
|
|
av_reduce(&num, &den, avctx->sample_aspect_ratio.num,
|
|
avctx->sample_aspect_ratio.den, 65535);
|
|
for (i = 0; i < FF_ARRAY_ELEMS(sar_idc); i++) {
|
|
if (num == sar_idc[i].num &&
|
|
den == sar_idc[i].den) {
|
|
vui->aspect_ratio_idc = i;
|
|
break;
|
|
}
|
|
}
|
|
if (i >= FF_ARRAY_ELEMS(sar_idc)) {
|
|
vui->aspect_ratio_idc = 255;
|
|
vui->sar_width = num;
|
|
vui->sar_height = den;
|
|
}
|
|
vui->aspect_ratio_info_present_flag = 1;
|
|
}
|
|
|
|
// Unspecified video format, from table E-2.
|
|
vui->video_format = 5;
|
|
vui->video_full_range_flag =
|
|
avctx->color_range == AVCOL_RANGE_JPEG;
|
|
vui->colour_primaries = avctx->color_primaries;
|
|
vui->transfer_characteristics = avctx->color_trc;
|
|
vui->matrix_coefficients = avctx->colorspace;
|
|
if (avctx->color_primaries != AVCOL_PRI_UNSPECIFIED ||
|
|
avctx->color_trc != AVCOL_TRC_UNSPECIFIED ||
|
|
avctx->colorspace != AVCOL_SPC_UNSPECIFIED)
|
|
vui->colour_description_present_flag = 1;
|
|
if (avctx->color_range != AVCOL_RANGE_UNSPECIFIED ||
|
|
vui->colour_description_present_flag)
|
|
vui->video_signal_type_present_flag = 1;
|
|
|
|
if (avctx->chroma_sample_location != AVCHROMA_LOC_UNSPECIFIED) {
|
|
vui->chroma_loc_info_present_flag = 1;
|
|
vui->chroma_sample_loc_type_top_field =
|
|
vui->chroma_sample_loc_type_bottom_field =
|
|
avctx->chroma_sample_location - 1;
|
|
}
|
|
|
|
vui->vui_timing_info_present_flag = 1;
|
|
vui->vui_num_units_in_tick = vps->vps_num_units_in_tick;
|
|
vui->vui_time_scale = vps->vps_time_scale;
|
|
vui->vui_poc_proportional_to_timing_flag = vps->vps_poc_proportional_to_timing_flag;
|
|
vui->vui_num_ticks_poc_diff_one_minus1 = vps->vps_num_ticks_poc_diff_one_minus1;
|
|
vui->vui_hrd_parameters_present_flag = 0;
|
|
|
|
vui->bitstream_restriction_flag = 1;
|
|
vui->motion_vectors_over_pic_boundaries_flag = 1;
|
|
vui->restricted_ref_pic_lists_flag = 1;
|
|
vui->max_bytes_per_pic_denom = 0;
|
|
vui->max_bits_per_min_cu_denom = 0;
|
|
vui->log2_max_mv_length_horizontal = 15;
|
|
vui->log2_max_mv_length_vertical = 15;
|
|
|
|
|
|
// PPS
|
|
|
|
pps->nal_unit_header = (H265RawNALUnitHeader) {
|
|
.nal_unit_type = HEVC_NAL_PPS,
|
|
.nuh_layer_id = 0,
|
|
.nuh_temporal_id_plus1 = 1,
|
|
};
|
|
|
|
pps->pps_pic_parameter_set_id = 0;
|
|
pps->pps_seq_parameter_set_id = sps->sps_seq_parameter_set_id;
|
|
|
|
pps->num_ref_idx_l0_default_active_minus1 = 0;
|
|
pps->num_ref_idx_l1_default_active_minus1 = 0;
|
|
|
|
pps->init_qp_minus26 = priv->fixed_qp_idr - 26;
|
|
|
|
pps->cu_qp_delta_enabled_flag = (ctx->va_rc_mode != VA_RC_CQP);
|
|
pps->diff_cu_qp_delta_depth = 0;
|
|
|
|
if (ctx->tile_rows && ctx->tile_cols) {
|
|
int uniform_spacing;
|
|
|
|
pps->tiles_enabled_flag = 1;
|
|
pps->num_tile_columns_minus1 = ctx->tile_cols - 1;
|
|
pps->num_tile_rows_minus1 = ctx->tile_rows - 1;
|
|
|
|
// Test whether the spacing provided matches the H.265 uniform
|
|
// spacing, and set the flag if it does.
|
|
uniform_spacing = 1;
|
|
for (i = 0; i <= pps->num_tile_columns_minus1 &&
|
|
uniform_spacing; i++) {
|
|
if (ctx->col_width[i] !=
|
|
(i + 1) * ctx->slice_block_cols / ctx->tile_cols -
|
|
i * ctx->slice_block_cols / ctx->tile_cols)
|
|
uniform_spacing = 0;
|
|
}
|
|
for (i = 0; i <= pps->num_tile_rows_minus1 &&
|
|
uniform_spacing; i++) {
|
|
if (ctx->row_height[i] !=
|
|
(i + 1) * ctx->slice_block_rows / ctx->tile_rows -
|
|
i * ctx->slice_block_rows / ctx->tile_rows)
|
|
uniform_spacing = 0;
|
|
}
|
|
pps->uniform_spacing_flag = uniform_spacing;
|
|
|
|
for (i = 0; i <= pps->num_tile_columns_minus1; i++)
|
|
pps->column_width_minus1[i] = ctx->col_width[i] - 1;
|
|
for (i = 0; i <= pps->num_tile_rows_minus1; i++)
|
|
pps->row_height_minus1[i] = ctx->row_height[i] - 1;
|
|
|
|
pps->loop_filter_across_tiles_enabled_flag = 1;
|
|
}
|
|
|
|
pps->pps_loop_filter_across_slices_enabled_flag = 1;
|
|
|
|
// Fill VAAPI parameter buffers.
|
|
|
|
*vseq = (VAEncSequenceParameterBufferHEVC) {
|
|
.general_profile_idc = vps->profile_tier_level.general_profile_idc,
|
|
.general_level_idc = vps->profile_tier_level.general_level_idc,
|
|
.general_tier_flag = vps->profile_tier_level.general_tier_flag,
|
|
|
|
.intra_period = ctx->gop_size,
|
|
.intra_idr_period = ctx->gop_size,
|
|
.ip_period = ctx->b_per_p + 1,
|
|
.bits_per_second = ctx->va_bit_rate,
|
|
|
|
.pic_width_in_luma_samples = sps->pic_width_in_luma_samples,
|
|
.pic_height_in_luma_samples = sps->pic_height_in_luma_samples,
|
|
|
|
.seq_fields.bits = {
|
|
.chroma_format_idc = sps->chroma_format_idc,
|
|
.separate_colour_plane_flag = sps->separate_colour_plane_flag,
|
|
.bit_depth_luma_minus8 = sps->bit_depth_luma_minus8,
|
|
.bit_depth_chroma_minus8 = sps->bit_depth_chroma_minus8,
|
|
.scaling_list_enabled_flag = sps->scaling_list_enabled_flag,
|
|
.strong_intra_smoothing_enabled_flag =
|
|
sps->strong_intra_smoothing_enabled_flag,
|
|
.amp_enabled_flag = sps->amp_enabled_flag,
|
|
.sample_adaptive_offset_enabled_flag =
|
|
sps->sample_adaptive_offset_enabled_flag,
|
|
.pcm_enabled_flag = sps->pcm_enabled_flag,
|
|
.pcm_loop_filter_disabled_flag = sps->pcm_loop_filter_disabled_flag,
|
|
.sps_temporal_mvp_enabled_flag = sps->sps_temporal_mvp_enabled_flag,
|
|
},
|
|
|
|
.log2_min_luma_coding_block_size_minus3 =
|
|
sps->log2_min_luma_coding_block_size_minus3,
|
|
.log2_diff_max_min_luma_coding_block_size =
|
|
sps->log2_diff_max_min_luma_coding_block_size,
|
|
.log2_min_transform_block_size_minus2 =
|
|
sps->log2_min_luma_transform_block_size_minus2,
|
|
.log2_diff_max_min_transform_block_size =
|
|
sps->log2_diff_max_min_luma_transform_block_size,
|
|
.max_transform_hierarchy_depth_inter =
|
|
sps->max_transform_hierarchy_depth_inter,
|
|
.max_transform_hierarchy_depth_intra =
|
|
sps->max_transform_hierarchy_depth_intra,
|
|
|
|
.pcm_sample_bit_depth_luma_minus1 =
|
|
sps->pcm_sample_bit_depth_luma_minus1,
|
|
.pcm_sample_bit_depth_chroma_minus1 =
|
|
sps->pcm_sample_bit_depth_chroma_minus1,
|
|
.log2_min_pcm_luma_coding_block_size_minus3 =
|
|
sps->log2_min_pcm_luma_coding_block_size_minus3,
|
|
.log2_max_pcm_luma_coding_block_size_minus3 =
|
|
sps->log2_min_pcm_luma_coding_block_size_minus3 +
|
|
sps->log2_diff_max_min_pcm_luma_coding_block_size,
|
|
|
|
.vui_parameters_present_flag = 0,
|
|
};
|
|
|
|
*vpic = (VAEncPictureParameterBufferHEVC) {
|
|
.decoded_curr_pic = {
|
|
.picture_id = VA_INVALID_ID,
|
|
.flags = VA_PICTURE_HEVC_INVALID,
|
|
},
|
|
|
|
.coded_buf = VA_INVALID_ID,
|
|
|
|
.collocated_ref_pic_index = 0xff,
|
|
|
|
.last_picture = 0,
|
|
|
|
.pic_init_qp = pps->init_qp_minus26 + 26,
|
|
.diff_cu_qp_delta_depth = pps->diff_cu_qp_delta_depth,
|
|
.pps_cb_qp_offset = pps->pps_cb_qp_offset,
|
|
.pps_cr_qp_offset = pps->pps_cr_qp_offset,
|
|
|
|
.num_tile_columns_minus1 = pps->num_tile_columns_minus1,
|
|
.num_tile_rows_minus1 = pps->num_tile_rows_minus1,
|
|
|
|
.log2_parallel_merge_level_minus2 = pps->log2_parallel_merge_level_minus2,
|
|
.ctu_max_bitsize_allowed = 0,
|
|
|
|
.num_ref_idx_l0_default_active_minus1 =
|
|
pps->num_ref_idx_l0_default_active_minus1,
|
|
.num_ref_idx_l1_default_active_minus1 =
|
|
pps->num_ref_idx_l1_default_active_minus1,
|
|
|
|
.slice_pic_parameter_set_id = pps->pps_pic_parameter_set_id,
|
|
|
|
.pic_fields.bits = {
|
|
.sign_data_hiding_enabled_flag = pps->sign_data_hiding_enabled_flag,
|
|
.constrained_intra_pred_flag = pps->constrained_intra_pred_flag,
|
|
.transform_skip_enabled_flag = pps->transform_skip_enabled_flag,
|
|
.cu_qp_delta_enabled_flag = pps->cu_qp_delta_enabled_flag,
|
|
.weighted_pred_flag = pps->weighted_pred_flag,
|
|
.weighted_bipred_flag = pps->weighted_bipred_flag,
|
|
.transquant_bypass_enabled_flag = pps->transquant_bypass_enabled_flag,
|
|
.tiles_enabled_flag = pps->tiles_enabled_flag,
|
|
.entropy_coding_sync_enabled_flag = pps->entropy_coding_sync_enabled_flag,
|
|
.loop_filter_across_tiles_enabled_flag =
|
|
pps->loop_filter_across_tiles_enabled_flag,
|
|
.scaling_list_data_present_flag = (sps->sps_scaling_list_data_present_flag |
|
|
pps->pps_scaling_list_data_present_flag),
|
|
.screen_content_flag = 0,
|
|
.enable_gpu_weighted_prediction = 0,
|
|
.no_output_of_prior_pics_flag = 0,
|
|
},
|
|
};
|
|
|
|
if (pps->tiles_enabled_flag) {
|
|
for (i = 0; i <= vpic->num_tile_rows_minus1; i++)
|
|
vpic->row_height_minus1[i] = pps->row_height_minus1[i];
|
|
for (i = 0; i <= vpic->num_tile_columns_minus1; i++)
|
|
vpic->column_width_minus1[i] = pps->column_width_minus1[i];
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vaapi_encode_h265_init_picture_params(AVCodecContext *avctx,
|
|
VAAPIEncodePicture *pic)
|
|
{
|
|
VAAPIEncodeContext *ctx = avctx->priv_data;
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
VAAPIEncodeH265Picture *hpic = pic->priv_data;
|
|
VAAPIEncodePicture *prev = pic->prev;
|
|
VAAPIEncodeH265Picture *hprev = prev ? prev->priv_data : NULL;
|
|
VAEncPictureParameterBufferHEVC *vpic = pic->codec_picture_params;
|
|
int i;
|
|
|
|
if (pic->type == PICTURE_TYPE_IDR) {
|
|
av_assert0(pic->display_order == pic->encode_order);
|
|
|
|
hpic->last_idr_frame = pic->display_order;
|
|
|
|
hpic->slice_nal_unit = HEVC_NAL_IDR_W_RADL;
|
|
hpic->slice_type = HEVC_SLICE_I;
|
|
hpic->pic_type = 0;
|
|
} else {
|
|
av_assert0(prev);
|
|
hpic->last_idr_frame = hprev->last_idr_frame;
|
|
|
|
if (pic->type == PICTURE_TYPE_I) {
|
|
hpic->slice_nal_unit = HEVC_NAL_CRA_NUT;
|
|
hpic->slice_type = HEVC_SLICE_I;
|
|
hpic->pic_type = 0;
|
|
} else if (pic->type == PICTURE_TYPE_P) {
|
|
av_assert0(pic->refs[0]);
|
|
hpic->slice_nal_unit = HEVC_NAL_TRAIL_R;
|
|
hpic->slice_type = HEVC_SLICE_P;
|
|
hpic->pic_type = 1;
|
|
} else {
|
|
VAAPIEncodePicture *irap_ref;
|
|
av_assert0(pic->refs[0] && pic->refs[1]);
|
|
for (irap_ref = pic; irap_ref; irap_ref = irap_ref->refs[1]) {
|
|
if (irap_ref->type == PICTURE_TYPE_I)
|
|
break;
|
|
}
|
|
if (pic->b_depth == ctx->max_b_depth) {
|
|
hpic->slice_nal_unit = irap_ref ? HEVC_NAL_RASL_N
|
|
: HEVC_NAL_TRAIL_N;
|
|
} else {
|
|
hpic->slice_nal_unit = irap_ref ? HEVC_NAL_RASL_R
|
|
: HEVC_NAL_TRAIL_R;
|
|
}
|
|
hpic->slice_type = HEVC_SLICE_B;
|
|
hpic->pic_type = 2;
|
|
}
|
|
}
|
|
hpic->pic_order_cnt = pic->display_order - hpic->last_idr_frame;
|
|
|
|
if (priv->aud) {
|
|
priv->aud_needed = 1;
|
|
priv->raw_aud = (H265RawAUD) {
|
|
.nal_unit_header = {
|
|
.nal_unit_type = HEVC_NAL_AUD,
|
|
.nuh_layer_id = 0,
|
|
.nuh_temporal_id_plus1 = 1,
|
|
},
|
|
.pic_type = hpic->pic_type,
|
|
};
|
|
} else {
|
|
priv->aud_needed = 0;
|
|
}
|
|
|
|
priv->sei_needed = 0;
|
|
|
|
// Only look for the metadata on I/IDR frame on the output. We
|
|
// may force an IDR frame on the output where the medadata gets
|
|
// changed on the input frame.
|
|
if ((priv->sei & SEI_MASTERING_DISPLAY) &&
|
|
(pic->type == PICTURE_TYPE_I || pic->type == PICTURE_TYPE_IDR)) {
|
|
AVFrameSideData *sd =
|
|
av_frame_get_side_data(pic->input_image,
|
|
AV_FRAME_DATA_MASTERING_DISPLAY_METADATA);
|
|
|
|
if (sd) {
|
|
AVMasteringDisplayMetadata *mdm =
|
|
(AVMasteringDisplayMetadata *)sd->data;
|
|
|
|
// SEI is needed when both the primaries and luminance are set
|
|
if (mdm->has_primaries && mdm->has_luminance) {
|
|
SEIRawMasteringDisplayColourVolume *mdcv =
|
|
&priv->sei_mastering_display;
|
|
const int mapping[3] = {1, 2, 0};
|
|
const int chroma_den = 50000;
|
|
const int luma_den = 10000;
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
const int j = mapping[i];
|
|
mdcv->display_primaries_x[i] =
|
|
FFMIN(lrint(chroma_den *
|
|
av_q2d(mdm->display_primaries[j][0])),
|
|
chroma_den);
|
|
mdcv->display_primaries_y[i] =
|
|
FFMIN(lrint(chroma_den *
|
|
av_q2d(mdm->display_primaries[j][1])),
|
|
chroma_den);
|
|
}
|
|
|
|
mdcv->white_point_x =
|
|
FFMIN(lrint(chroma_den * av_q2d(mdm->white_point[0])),
|
|
chroma_den);
|
|
mdcv->white_point_y =
|
|
FFMIN(lrint(chroma_den * av_q2d(mdm->white_point[1])),
|
|
chroma_den);
|
|
|
|
mdcv->max_display_mastering_luminance =
|
|
lrint(luma_den * av_q2d(mdm->max_luminance));
|
|
mdcv->min_display_mastering_luminance =
|
|
FFMIN(lrint(luma_den * av_q2d(mdm->min_luminance)),
|
|
mdcv->max_display_mastering_luminance);
|
|
|
|
priv->sei_needed |= SEI_MASTERING_DISPLAY;
|
|
}
|
|
}
|
|
}
|
|
|
|
if ((priv->sei & SEI_CONTENT_LIGHT_LEVEL) &&
|
|
(pic->type == PICTURE_TYPE_I || pic->type == PICTURE_TYPE_IDR)) {
|
|
AVFrameSideData *sd =
|
|
av_frame_get_side_data(pic->input_image,
|
|
AV_FRAME_DATA_CONTENT_LIGHT_LEVEL);
|
|
|
|
if (sd) {
|
|
AVContentLightMetadata *clm =
|
|
(AVContentLightMetadata *)sd->data;
|
|
SEIRawContentLightLevelInfo *clli =
|
|
&priv->sei_content_light_level;
|
|
|
|
clli->max_content_light_level = FFMIN(clm->MaxCLL, 65535);
|
|
clli->max_pic_average_light_level = FFMIN(clm->MaxFALL, 65535);
|
|
|
|
priv->sei_needed |= SEI_CONTENT_LIGHT_LEVEL;
|
|
}
|
|
}
|
|
|
|
vpic->decoded_curr_pic = (VAPictureHEVC) {
|
|
.picture_id = pic->recon_surface,
|
|
.pic_order_cnt = hpic->pic_order_cnt,
|
|
.flags = 0,
|
|
};
|
|
|
|
for (i = 0; i < pic->nb_refs; i++) {
|
|
VAAPIEncodePicture *ref = pic->refs[i];
|
|
VAAPIEncodeH265Picture *href;
|
|
|
|
av_assert0(ref && ref->encode_order < pic->encode_order);
|
|
href = ref->priv_data;
|
|
|
|
vpic->reference_frames[i] = (VAPictureHEVC) {
|
|
.picture_id = ref->recon_surface,
|
|
.pic_order_cnt = href->pic_order_cnt,
|
|
.flags = (ref->display_order < pic->display_order ?
|
|
VA_PICTURE_HEVC_RPS_ST_CURR_BEFORE : 0) |
|
|
(ref->display_order > pic->display_order ?
|
|
VA_PICTURE_HEVC_RPS_ST_CURR_AFTER : 0),
|
|
};
|
|
}
|
|
for (; i < FF_ARRAY_ELEMS(vpic->reference_frames); i++) {
|
|
vpic->reference_frames[i] = (VAPictureHEVC) {
|
|
.picture_id = VA_INVALID_ID,
|
|
.flags = VA_PICTURE_HEVC_INVALID,
|
|
};
|
|
}
|
|
|
|
vpic->coded_buf = pic->output_buffer;
|
|
|
|
vpic->nal_unit_type = hpic->slice_nal_unit;
|
|
|
|
switch (pic->type) {
|
|
case PICTURE_TYPE_IDR:
|
|
vpic->pic_fields.bits.idr_pic_flag = 1;
|
|
vpic->pic_fields.bits.coding_type = 1;
|
|
vpic->pic_fields.bits.reference_pic_flag = 1;
|
|
break;
|
|
case PICTURE_TYPE_I:
|
|
vpic->pic_fields.bits.idr_pic_flag = 0;
|
|
vpic->pic_fields.bits.coding_type = 1;
|
|
vpic->pic_fields.bits.reference_pic_flag = 1;
|
|
break;
|
|
case PICTURE_TYPE_P:
|
|
vpic->pic_fields.bits.idr_pic_flag = 0;
|
|
vpic->pic_fields.bits.coding_type = 2;
|
|
vpic->pic_fields.bits.reference_pic_flag = 1;
|
|
break;
|
|
case PICTURE_TYPE_B:
|
|
vpic->pic_fields.bits.idr_pic_flag = 0;
|
|
vpic->pic_fields.bits.coding_type = 3;
|
|
vpic->pic_fields.bits.reference_pic_flag = 0;
|
|
break;
|
|
default:
|
|
av_assert0(0 && "invalid picture type");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vaapi_encode_h265_init_slice_params(AVCodecContext *avctx,
|
|
VAAPIEncodePicture *pic,
|
|
VAAPIEncodeSlice *slice)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
VAAPIEncodeH265Picture *hpic = pic->priv_data;
|
|
const H265RawSPS *sps = &priv->raw_sps;
|
|
const H265RawPPS *pps = &priv->raw_pps;
|
|
H265RawSliceHeader *sh = &priv->raw_slice.header;
|
|
VAEncPictureParameterBufferHEVC *vpic = pic->codec_picture_params;
|
|
VAEncSliceParameterBufferHEVC *vslice = slice->codec_slice_params;
|
|
int i;
|
|
|
|
sh->nal_unit_header = (H265RawNALUnitHeader) {
|
|
.nal_unit_type = hpic->slice_nal_unit,
|
|
.nuh_layer_id = 0,
|
|
.nuh_temporal_id_plus1 = 1,
|
|
};
|
|
|
|
sh->slice_pic_parameter_set_id = pps->pps_pic_parameter_set_id;
|
|
|
|
sh->first_slice_segment_in_pic_flag = slice->index == 0;
|
|
sh->slice_segment_address = slice->block_start;
|
|
|
|
sh->slice_type = hpic->slice_type;
|
|
|
|
sh->slice_pic_order_cnt_lsb = hpic->pic_order_cnt &
|
|
(1 << (sps->log2_max_pic_order_cnt_lsb_minus4 + 4)) - 1;
|
|
|
|
if (pic->type != PICTURE_TYPE_IDR) {
|
|
H265RawSTRefPicSet *rps;
|
|
const VAAPIEncodeH265Picture *strp;
|
|
int rps_poc[MAX_DPB_SIZE];
|
|
int rps_used[MAX_DPB_SIZE];
|
|
int i, j, poc, rps_pics;
|
|
|
|
sh->short_term_ref_pic_set_sps_flag = 0;
|
|
|
|
rps = &sh->short_term_ref_pic_set;
|
|
memset(rps, 0, sizeof(*rps));
|
|
|
|
rps_pics = 0;
|
|
for (i = 0; i < pic->nb_refs; i++) {
|
|
strp = pic->refs[i]->priv_data;
|
|
rps_poc[rps_pics] = strp->pic_order_cnt;
|
|
rps_used[rps_pics] = 1;
|
|
++rps_pics;
|
|
}
|
|
for (i = 0; i < pic->nb_dpb_pics; i++) {
|
|
if (pic->dpb[i] == pic)
|
|
continue;
|
|
for (j = 0; j < pic->nb_refs; j++) {
|
|
if (pic->dpb[i] == pic->refs[j])
|
|
break;
|
|
}
|
|
if (j < pic->nb_refs)
|
|
continue;
|
|
strp = pic->dpb[i]->priv_data;
|
|
rps_poc[rps_pics] = strp->pic_order_cnt;
|
|
rps_used[rps_pics] = 0;
|
|
++rps_pics;
|
|
}
|
|
|
|
for (i = 1; i < rps_pics; i++) {
|
|
for (j = i; j > 0; j--) {
|
|
if (rps_poc[j] > rps_poc[j - 1])
|
|
break;
|
|
av_assert0(rps_poc[j] != rps_poc[j - 1]);
|
|
FFSWAP(int, rps_poc[j], rps_poc[j - 1]);
|
|
FFSWAP(int, rps_used[j], rps_used[j - 1]);
|
|
}
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_DEBUG, "RPS for POC %d:",
|
|
hpic->pic_order_cnt);
|
|
for (i = 0; i < rps_pics; i++) {
|
|
av_log(avctx, AV_LOG_DEBUG, " (%d,%d)",
|
|
rps_poc[i], rps_used[i]);
|
|
}
|
|
av_log(avctx, AV_LOG_DEBUG, "\n");
|
|
|
|
for (i = 0; i < rps_pics; i++) {
|
|
av_assert0(rps_poc[i] != hpic->pic_order_cnt);
|
|
if (rps_poc[i] > hpic->pic_order_cnt)
|
|
break;
|
|
}
|
|
|
|
rps->num_negative_pics = i;
|
|
poc = hpic->pic_order_cnt;
|
|
for (j = i - 1; j >= 0; j--) {
|
|
rps->delta_poc_s0_minus1[i - 1 - j] = poc - rps_poc[j] - 1;
|
|
rps->used_by_curr_pic_s0_flag[i - 1 - j] = rps_used[j];
|
|
poc = rps_poc[j];
|
|
}
|
|
|
|
rps->num_positive_pics = rps_pics - i;
|
|
poc = hpic->pic_order_cnt;
|
|
for (j = i; j < rps_pics; j++) {
|
|
rps->delta_poc_s1_minus1[j - i] = rps_poc[j] - poc - 1;
|
|
rps->used_by_curr_pic_s1_flag[j - i] = rps_used[j];
|
|
poc = rps_poc[j];
|
|
}
|
|
|
|
sh->num_long_term_sps = 0;
|
|
sh->num_long_term_pics = 0;
|
|
|
|
sh->slice_temporal_mvp_enabled_flag =
|
|
sps->sps_temporal_mvp_enabled_flag;
|
|
if (sh->slice_temporal_mvp_enabled_flag) {
|
|
sh->collocated_from_l0_flag = sh->slice_type == HEVC_SLICE_B;
|
|
sh->collocated_ref_idx = 0;
|
|
}
|
|
|
|
sh->num_ref_idx_active_override_flag = 0;
|
|
sh->num_ref_idx_l0_active_minus1 = pps->num_ref_idx_l0_default_active_minus1;
|
|
sh->num_ref_idx_l1_active_minus1 = pps->num_ref_idx_l1_default_active_minus1;
|
|
}
|
|
|
|
sh->slice_sao_luma_flag = sh->slice_sao_chroma_flag =
|
|
sps->sample_adaptive_offset_enabled_flag;
|
|
|
|
if (pic->type == PICTURE_TYPE_B)
|
|
sh->slice_qp_delta = priv->fixed_qp_b - (pps->init_qp_minus26 + 26);
|
|
else if (pic->type == PICTURE_TYPE_P)
|
|
sh->slice_qp_delta = priv->fixed_qp_p - (pps->init_qp_minus26 + 26);
|
|
else
|
|
sh->slice_qp_delta = priv->fixed_qp_idr - (pps->init_qp_minus26 + 26);
|
|
|
|
|
|
*vslice = (VAEncSliceParameterBufferHEVC) {
|
|
.slice_segment_address = sh->slice_segment_address,
|
|
.num_ctu_in_slice = slice->block_size,
|
|
|
|
.slice_type = sh->slice_type,
|
|
.slice_pic_parameter_set_id = sh->slice_pic_parameter_set_id,
|
|
|
|
.num_ref_idx_l0_active_minus1 = sh->num_ref_idx_l0_active_minus1,
|
|
.num_ref_idx_l1_active_minus1 = sh->num_ref_idx_l1_active_minus1,
|
|
|
|
.luma_log2_weight_denom = sh->luma_log2_weight_denom,
|
|
.delta_chroma_log2_weight_denom = sh->delta_chroma_log2_weight_denom,
|
|
|
|
.max_num_merge_cand = 5 - sh->five_minus_max_num_merge_cand,
|
|
|
|
.slice_qp_delta = sh->slice_qp_delta,
|
|
.slice_cb_qp_offset = sh->slice_cb_qp_offset,
|
|
.slice_cr_qp_offset = sh->slice_cr_qp_offset,
|
|
|
|
.slice_beta_offset_div2 = sh->slice_beta_offset_div2,
|
|
.slice_tc_offset_div2 = sh->slice_tc_offset_div2,
|
|
|
|
.slice_fields.bits = {
|
|
.last_slice_of_pic_flag = slice->index == pic->nb_slices - 1,
|
|
.dependent_slice_segment_flag = sh->dependent_slice_segment_flag,
|
|
.colour_plane_id = sh->colour_plane_id,
|
|
.slice_temporal_mvp_enabled_flag =
|
|
sh->slice_temporal_mvp_enabled_flag,
|
|
.slice_sao_luma_flag = sh->slice_sao_luma_flag,
|
|
.slice_sao_chroma_flag = sh->slice_sao_chroma_flag,
|
|
.num_ref_idx_active_override_flag =
|
|
sh->num_ref_idx_active_override_flag,
|
|
.mvd_l1_zero_flag = sh->mvd_l1_zero_flag,
|
|
.cabac_init_flag = sh->cabac_init_flag,
|
|
.slice_deblocking_filter_disabled_flag =
|
|
sh->slice_deblocking_filter_disabled_flag,
|
|
.slice_loop_filter_across_slices_enabled_flag =
|
|
sh->slice_loop_filter_across_slices_enabled_flag,
|
|
.collocated_from_l0_flag = sh->collocated_from_l0_flag,
|
|
},
|
|
};
|
|
|
|
for (i = 0; i < FF_ARRAY_ELEMS(vslice->ref_pic_list0); i++) {
|
|
vslice->ref_pic_list0[i].picture_id = VA_INVALID_ID;
|
|
vslice->ref_pic_list0[i].flags = VA_PICTURE_HEVC_INVALID;
|
|
vslice->ref_pic_list1[i].picture_id = VA_INVALID_ID;
|
|
vslice->ref_pic_list1[i].flags = VA_PICTURE_HEVC_INVALID;
|
|
}
|
|
|
|
av_assert0(pic->nb_refs <= 2);
|
|
if (pic->nb_refs >= 1) {
|
|
// Backward reference for P- or B-frame.
|
|
av_assert0(pic->type == PICTURE_TYPE_P ||
|
|
pic->type == PICTURE_TYPE_B);
|
|
vslice->ref_pic_list0[0] = vpic->reference_frames[0];
|
|
}
|
|
if (pic->nb_refs >= 2) {
|
|
// Forward reference for B-frame.
|
|
av_assert0(pic->type == PICTURE_TYPE_B);
|
|
vslice->ref_pic_list1[0] = vpic->reference_frames[1];
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int vaapi_encode_h265_configure(AVCodecContext *avctx)
|
|
{
|
|
VAAPIEncodeContext *ctx = avctx->priv_data;
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
int err;
|
|
|
|
err = ff_cbs_init(&priv->cbc, AV_CODEC_ID_HEVC, avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
if (ctx->va_rc_mode == VA_RC_CQP) {
|
|
// Note that VAAPI only supports positive QP values - the range is
|
|
// therefore always bounded below by 1, even in 10-bit mode where
|
|
// it should go down to -12.
|
|
|
|
priv->fixed_qp_p = av_clip(ctx->rc_quality, 1, 51);
|
|
if (avctx->i_quant_factor > 0.0)
|
|
priv->fixed_qp_idr =
|
|
av_clip((avctx->i_quant_factor * priv->fixed_qp_p +
|
|
avctx->i_quant_offset) + 0.5, 1, 51);
|
|
else
|
|
priv->fixed_qp_idr = priv->fixed_qp_p;
|
|
if (avctx->b_quant_factor > 0.0)
|
|
priv->fixed_qp_b =
|
|
av_clip((avctx->b_quant_factor * priv->fixed_qp_p +
|
|
avctx->b_quant_offset) + 0.5, 1, 51);
|
|
else
|
|
priv->fixed_qp_b = priv->fixed_qp_p;
|
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Using fixed QP = "
|
|
"%d / %d / %d for IDR- / P- / B-frames.\n",
|
|
priv->fixed_qp_idr, priv->fixed_qp_p, priv->fixed_qp_b);
|
|
|
|
} else {
|
|
// These still need to be set for init_qp/slice_qp_delta.
|
|
priv->fixed_qp_idr = 30;
|
|
priv->fixed_qp_p = 30;
|
|
priv->fixed_qp_b = 30;
|
|
}
|
|
|
|
ctx->roi_quant_range = 51 + 6 * (ctx->profile->depth - 8);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const VAAPIEncodeProfile vaapi_encode_h265_profiles[] = {
|
|
{ FF_PROFILE_HEVC_MAIN, 8, 3, 1, 1, VAProfileHEVCMain },
|
|
{ FF_PROFILE_HEVC_REXT, 8, 3, 1, 1, VAProfileHEVCMain },
|
|
#if VA_CHECK_VERSION(0, 37, 0)
|
|
{ FF_PROFILE_HEVC_MAIN_10, 10, 3, 1, 1, VAProfileHEVCMain10 },
|
|
{ FF_PROFILE_HEVC_REXT, 10, 3, 1, 1, VAProfileHEVCMain10 },
|
|
#endif
|
|
#if VA_CHECK_VERSION(1, 2, 0)
|
|
{ FF_PROFILE_HEVC_REXT, 8, 3, 1, 0, VAProfileHEVCMain422_10 },
|
|
{ FF_PROFILE_HEVC_REXT, 10, 3, 1, 0, VAProfileHEVCMain422_10 },
|
|
#endif
|
|
{ FF_PROFILE_UNKNOWN }
|
|
};
|
|
|
|
static const VAAPIEncodeType vaapi_encode_type_h265 = {
|
|
.profiles = vaapi_encode_h265_profiles,
|
|
|
|
.flags = FLAG_SLICE_CONTROL |
|
|
FLAG_B_PICTURES |
|
|
FLAG_B_PICTURE_REFERENCES |
|
|
FLAG_NON_IDR_KEY_PICTURES,
|
|
|
|
.default_quality = 25,
|
|
|
|
.configure = &vaapi_encode_h265_configure,
|
|
|
|
.picture_priv_data_size = sizeof(VAAPIEncodeH265Picture),
|
|
|
|
.sequence_params_size = sizeof(VAEncSequenceParameterBufferHEVC),
|
|
.init_sequence_params = &vaapi_encode_h265_init_sequence_params,
|
|
|
|
.picture_params_size = sizeof(VAEncPictureParameterBufferHEVC),
|
|
.init_picture_params = &vaapi_encode_h265_init_picture_params,
|
|
|
|
.slice_params_size = sizeof(VAEncSliceParameterBufferHEVC),
|
|
.init_slice_params = &vaapi_encode_h265_init_slice_params,
|
|
|
|
.sequence_header_type = VAEncPackedHeaderSequence,
|
|
.write_sequence_header = &vaapi_encode_h265_write_sequence_header,
|
|
|
|
.slice_header_type = VAEncPackedHeaderHEVC_Slice,
|
|
.write_slice_header = &vaapi_encode_h265_write_slice_header,
|
|
|
|
.write_extra_header = &vaapi_encode_h265_write_extra_header,
|
|
};
|
|
|
|
static av_cold int vaapi_encode_h265_init(AVCodecContext *avctx)
|
|
{
|
|
VAAPIEncodeContext *ctx = avctx->priv_data;
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
|
|
ctx->codec = &vaapi_encode_type_h265;
|
|
|
|
if (avctx->profile == FF_PROFILE_UNKNOWN)
|
|
avctx->profile = priv->profile;
|
|
if (avctx->level == FF_LEVEL_UNKNOWN)
|
|
avctx->level = priv->level;
|
|
|
|
if (avctx->level != FF_LEVEL_UNKNOWN && avctx->level & ~0xff) {
|
|
av_log(avctx, AV_LOG_ERROR, "Invalid level %d: must fit "
|
|
"in 8-bit unsigned integer.\n", avctx->level);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
ctx->desired_packed_headers =
|
|
VA_ENC_PACKED_HEADER_SEQUENCE | // VPS, SPS and PPS.
|
|
VA_ENC_PACKED_HEADER_SLICE | // Slice headers.
|
|
VA_ENC_PACKED_HEADER_MISC; // SEI
|
|
|
|
ctx->surface_width = FFALIGN(avctx->width, 16);
|
|
ctx->surface_height = FFALIGN(avctx->height, 16);
|
|
|
|
// CTU size is currently hard-coded to 32.
|
|
ctx->slice_block_width = ctx->slice_block_height = 32;
|
|
|
|
if (priv->qp > 0)
|
|
ctx->explicit_qp = priv->qp;
|
|
|
|
return ff_vaapi_encode_init(avctx);
|
|
}
|
|
|
|
static av_cold int vaapi_encode_h265_close(AVCodecContext *avctx)
|
|
{
|
|
VAAPIEncodeH265Context *priv = avctx->priv_data;
|
|
|
|
ff_cbs_fragment_free(&priv->current_access_unit);
|
|
ff_cbs_close(&priv->cbc);
|
|
|
|
return ff_vaapi_encode_close(avctx);
|
|
}
|
|
|
|
#define OFFSET(x) offsetof(VAAPIEncodeH265Context, x)
|
|
#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM)
|
|
static const AVOption vaapi_encode_h265_options[] = {
|
|
VAAPI_ENCODE_COMMON_OPTIONS,
|
|
VAAPI_ENCODE_RC_OPTIONS,
|
|
|
|
{ "qp", "Constant QP (for P-frames; scaled by qfactor/qoffset for I/B)",
|
|
OFFSET(qp), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 52, FLAGS },
|
|
|
|
{ "aud", "Include AUD",
|
|
OFFSET(aud), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS },
|
|
|
|
{ "profile", "Set profile (general_profile_idc)",
|
|
OFFSET(profile), AV_OPT_TYPE_INT,
|
|
{ .i64 = FF_PROFILE_UNKNOWN }, FF_PROFILE_UNKNOWN, 0xff, FLAGS, "profile" },
|
|
|
|
#define PROFILE(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
|
|
{ .i64 = value }, 0, 0, FLAGS, "profile"
|
|
{ PROFILE("main", FF_PROFILE_HEVC_MAIN) },
|
|
{ PROFILE("main10", FF_PROFILE_HEVC_MAIN_10) },
|
|
{ PROFILE("rext", FF_PROFILE_HEVC_REXT) },
|
|
#undef PROFILE
|
|
|
|
{ "tier", "Set tier (general_tier_flag)",
|
|
OFFSET(tier), AV_OPT_TYPE_INT,
|
|
{ .i64 = 0 }, 0, 1, FLAGS, "tier" },
|
|
{ "main", NULL, 0, AV_OPT_TYPE_CONST,
|
|
{ .i64 = 0 }, 0, 0, FLAGS, "tier" },
|
|
{ "high", NULL, 0, AV_OPT_TYPE_CONST,
|
|
{ .i64 = 1 }, 0, 0, FLAGS, "tier" },
|
|
|
|
{ "level", "Set level (general_level_idc)",
|
|
OFFSET(level), AV_OPT_TYPE_INT,
|
|
{ .i64 = FF_LEVEL_UNKNOWN }, FF_LEVEL_UNKNOWN, 0xff, FLAGS, "level" },
|
|
|
|
#define LEVEL(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
|
|
{ .i64 = value }, 0, 0, FLAGS, "level"
|
|
{ LEVEL("1", 30) },
|
|
{ LEVEL("2", 60) },
|
|
{ LEVEL("2.1", 63) },
|
|
{ LEVEL("3", 90) },
|
|
{ LEVEL("3.1", 93) },
|
|
{ LEVEL("4", 120) },
|
|
{ LEVEL("4.1", 123) },
|
|
{ LEVEL("5", 150) },
|
|
{ LEVEL("5.1", 153) },
|
|
{ LEVEL("5.2", 156) },
|
|
{ LEVEL("6", 180) },
|
|
{ LEVEL("6.1", 183) },
|
|
{ LEVEL("6.2", 186) },
|
|
#undef LEVEL
|
|
|
|
{ "sei", "Set SEI to include",
|
|
OFFSET(sei), AV_OPT_TYPE_FLAGS,
|
|
{ .i64 = SEI_MASTERING_DISPLAY | SEI_CONTENT_LIGHT_LEVEL },
|
|
0, INT_MAX, FLAGS, "sei" },
|
|
{ "hdr",
|
|
"Include HDR metadata for mastering display colour volume "
|
|
"and content light level information",
|
|
0, AV_OPT_TYPE_CONST,
|
|
{ .i64 = SEI_MASTERING_DISPLAY | SEI_CONTENT_LIGHT_LEVEL },
|
|
INT_MIN, INT_MAX, FLAGS, "sei" },
|
|
|
|
{ "tiles", "Tile columns x rows",
|
|
OFFSET(common.tile_cols), AV_OPT_TYPE_IMAGE_SIZE,
|
|
{ .str = NULL }, 0, 0, FLAGS },
|
|
|
|
{ NULL },
|
|
};
|
|
|
|
static const FFCodecDefault vaapi_encode_h265_defaults[] = {
|
|
{ "b", "0" },
|
|
{ "bf", "2" },
|
|
{ "g", "120" },
|
|
{ "i_qfactor", "1" },
|
|
{ "i_qoffset", "0" },
|
|
{ "b_qfactor", "6/5" },
|
|
{ "b_qoffset", "0" },
|
|
{ "qmin", "-1" },
|
|
{ "qmax", "-1" },
|
|
{ NULL },
|
|
};
|
|
|
|
static const AVClass vaapi_encode_h265_class = {
|
|
.class_name = "h265_vaapi",
|
|
.item_name = av_default_item_name,
|
|
.option = vaapi_encode_h265_options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
const FFCodec ff_hevc_vaapi_encoder = {
|
|
.p.name = "hevc_vaapi",
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("H.265/HEVC (VAAPI)"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_HEVC,
|
|
.priv_data_size = sizeof(VAAPIEncodeH265Context),
|
|
.init = &vaapi_encode_h265_init,
|
|
.receive_packet = &ff_vaapi_encode_receive_packet,
|
|
.close = &vaapi_encode_h265_close,
|
|
.p.priv_class = &vaapi_encode_h265_class,
|
|
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
|
AV_CODEC_CAP_DR1,
|
|
.caps_internal = FF_CODEC_CAP_INIT_CLEANUP,
|
|
.defaults = vaapi_encode_h265_defaults,
|
|
.p.pix_fmts = (const enum AVPixelFormat[]) {
|
|
AV_PIX_FMT_VAAPI,
|
|
AV_PIX_FMT_NONE,
|
|
},
|
|
.hw_configs = ff_vaapi_encode_hw_configs,
|
|
.p.wrapper_name = "vaapi",
|
|
};
|