You've already forked FFmpeg
mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2025-10-06 05:47:18 +02:00
1402 lines
54 KiB
C
1402 lines
54 KiB
C
/*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/mem.h"
|
|
|
|
#include "cbs.h"
|
|
#include "cbs_av1.h"
|
|
#include "av1_levels.h"
|
|
#include "libavutil/mastering_display_metadata.h"
|
|
|
|
#include "codec_internal.h"
|
|
#include "vulkan_encode.h"
|
|
|
|
#include "libavutil/avassert.h"
|
|
|
|
const FFVulkanEncodeDescriptor ff_vk_enc_av1_desc = {
|
|
.codec_id = AV_CODEC_ID_AV1,
|
|
.encode_extension = FF_VK_EXT_VIDEO_ENCODE_AV1,
|
|
.encode_op = VK_VIDEO_CODEC_OPERATION_ENCODE_AV1_BIT_KHR,
|
|
.ext_props = {
|
|
.extensionName = VK_STD_VULKAN_VIDEO_CODEC_AV1_ENCODE_EXTENSION_NAME,
|
|
.specVersion = VK_STD_VULKAN_VIDEO_CODEC_AV1_ENCODE_SPEC_VERSION,
|
|
},
|
|
};
|
|
|
|
enum UnitElems {
|
|
UNIT_MASTERING_DISPLAY = 1 << 0,
|
|
UNIT_CONTENT_LIGHT_LEVEL = 1 << 1,
|
|
};
|
|
|
|
typedef struct VulkanEncodeAV1Picture {
|
|
int slot;
|
|
int64_t last_idr_frame;
|
|
|
|
enum UnitElems units_needed;
|
|
|
|
StdVideoAV1TileInfo tile_info;
|
|
StdVideoAV1Quantization quantization;
|
|
StdVideoAV1Segmentation segmentation;
|
|
StdVideoAV1LoopFilter loop_filter;
|
|
StdVideoAV1CDEF cdef;
|
|
StdVideoAV1LoopRestoration loop_restoration;
|
|
StdVideoAV1GlobalMotion global_motion;
|
|
|
|
StdVideoEncodeAV1PictureInfo av1pic_info;
|
|
VkVideoEncodeAV1PictureInfoKHR vkav1pic_info;
|
|
|
|
StdVideoEncodeAV1ExtensionHeader ext_header;
|
|
StdVideoEncodeAV1ReferenceInfo av1dpb_info;
|
|
VkVideoEncodeAV1DpbSlotInfoKHR vkav1dpb_info;
|
|
|
|
VkVideoEncodeAV1RateControlInfoKHR vkrc_info;
|
|
VkVideoEncodeAV1RateControlLayerInfoKHR vkrc_layer_info;
|
|
VkVideoEncodeAV1GopRemainingFrameInfoKHR vkrc_remaining;
|
|
} VulkanEncodeAV1Picture;
|
|
|
|
typedef struct VulkanEncodeAV1Context {
|
|
FFVulkanEncodeContext common;
|
|
|
|
CodedBitstreamContext *cbs;
|
|
CodedBitstreamFragment current_access_unit;
|
|
|
|
enum UnitElems unit_elems;
|
|
AV1RawOBU seq_hdr_obu;
|
|
AV1RawOBU meta_cll_obu;
|
|
AV1RawOBU meta_mastering_obu;
|
|
|
|
VkVideoEncodeAV1ProfileInfoKHR profile;
|
|
|
|
VkVideoEncodeAV1CapabilitiesKHR caps;
|
|
VkVideoEncodeAV1QualityLevelPropertiesKHR quality_props;
|
|
|
|
uint64_t hrd_buffer_size;
|
|
uint64_t initial_buffer_fullness;
|
|
|
|
int uniform_tile;
|
|
int tile_cols;
|
|
int tile_rows;
|
|
|
|
int seq_tier;
|
|
int seq_level_idx;
|
|
|
|
int q_idx_idr;
|
|
int q_idx_p;
|
|
int q_idx_b;
|
|
|
|
uint8_t *padding_payload;
|
|
} VulkanEncodeAV1Context;
|
|
|
|
static int init_pic_rc(AVCodecContext *avctx, FFHWBaseEncodePicture *pic,
|
|
VkVideoEncodeRateControlInfoKHR *rc_info,
|
|
VkVideoEncodeRateControlLayerInfoKHR *rc_layer)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
VulkanEncodeAV1Picture *ap = pic->codec_priv;
|
|
|
|
/* This can be easy to calculate */
|
|
ap->vkrc_remaining = (VkVideoEncodeAV1GopRemainingFrameInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_GOP_REMAINING_FRAME_INFO_KHR,
|
|
.useGopRemainingFrames = 0,
|
|
.gopRemainingIntra = 0,
|
|
.gopRemainingPredictive = 0,
|
|
.gopRemainingBipredictive = 0,
|
|
};
|
|
|
|
ap->vkrc_info = (VkVideoEncodeAV1RateControlInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_RATE_CONTROL_INFO_KHR,
|
|
.flags = VK_VIDEO_ENCODE_AV1_RATE_CONTROL_REFERENCE_PATTERN_FLAT_BIT_KHR |
|
|
VK_VIDEO_ENCODE_AV1_RATE_CONTROL_REGULAR_GOP_BIT_KHR,
|
|
.gopFrameCount = ctx->base.gop_size,
|
|
.keyFramePeriod = ctx->base.gop_size,
|
|
.consecutiveBipredictiveFrameCount = FFMAX(ctx->base.b_per_p - 1, 0),
|
|
.temporalLayerCount = 0,
|
|
};
|
|
rc_info->pNext = &ap->vkrc_info;
|
|
|
|
if (rc_info->rateControlMode > VK_VIDEO_ENCODE_RATE_CONTROL_MODE_DISABLED_BIT_KHR) {
|
|
rc_info->virtualBufferSizeInMs = (enc->hrd_buffer_size * 1000LL) / avctx->bit_rate;
|
|
rc_info->initialVirtualBufferSizeInMs = (enc->initial_buffer_fullness * 1000LL) / avctx->bit_rate;
|
|
|
|
ap->vkrc_layer_info = (VkVideoEncodeAV1RateControlLayerInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_RATE_CONTROL_LAYER_INFO_KHR,
|
|
|
|
.useMinQIndex = avctx->qmin > 0,
|
|
.minQIndex.intraQIndex = avctx->qmin > 0 ? avctx->qmin : 0,
|
|
.minQIndex.predictiveQIndex = avctx->qmin > 0 ? avctx->qmin : 0,
|
|
.minQIndex.bipredictiveQIndex = avctx->qmin > 0 ? avctx->qmin : 0,
|
|
|
|
.useMaxQIndex = avctx->qmax > 0,
|
|
.maxQIndex.intraQIndex = avctx->qmax > 0 ? avctx->qmax : 0,
|
|
.maxQIndex.predictiveQIndex = avctx->qmax > 0 ? avctx->qmax : 0,
|
|
.maxQIndex.bipredictiveQIndex = avctx->qmax > 0 ? avctx->qmax : 0,
|
|
|
|
.useMaxFrameSize = 0,
|
|
};
|
|
rc_layer->pNext = &ap->vkrc_layer_info;
|
|
ap->vkrc_info.temporalLayerCount = 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void set_name_slot(int slot, int *slot_indices, uint32_t allowed_idx, int group)
|
|
{
|
|
int from = group ? AV1_REF_FRAME_GOLDEN : 0;
|
|
int to = group ? AV1_REFS_PER_FRAME : AV1_REF_FRAME_GOLDEN;
|
|
|
|
for (int i = from; i < to; i++) {
|
|
if ((slot_indices[i] == -1) && (allowed_idx & (1 << i))) {
|
|
slot_indices[i] = slot;
|
|
return;
|
|
}
|
|
}
|
|
|
|
av_assert0(0);
|
|
}
|
|
|
|
|
|
static int init_pic_params(AVCodecContext *avctx, FFHWBaseEncodePicture *pic,
|
|
VkVideoEncodeInfoKHR *encode_info)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFHWBaseEncodeContext *base_ctx = &ctx->base;
|
|
|
|
VulkanEncodeAV1Picture *ap = pic->codec_priv;
|
|
FFHWBaseEncodePicture *ref;
|
|
VulkanEncodeAV1Picture *ap_ref;
|
|
VkVideoReferenceSlotInfoKHR *ref_slot;
|
|
|
|
uint32_t ref_name_mask = 0x0;
|
|
int name_slots[STD_VIDEO_AV1_REFS_PER_FRAME];
|
|
|
|
StdVideoAV1Segmentation *segmentation = &ap->segmentation;
|
|
StdVideoAV1LoopFilter *loop_filter = &ap->loop_filter;
|
|
StdVideoAV1Quantization *quantization = &ap->quantization;
|
|
StdVideoAV1CDEF *cdef = &ap->cdef;
|
|
StdVideoAV1LoopRestoration *loop_restoration = &ap->loop_restoration;
|
|
StdVideoAV1GlobalMotion *global_motion = &ap->global_motion;
|
|
StdVideoAV1TileInfo *tile_info = &ap->tile_info;
|
|
static const int8_t default_loop_filter_ref_deltas[STD_VIDEO_AV1_TOTAL_REFS_PER_FRAME] =
|
|
{ 1, 0, 0, 0, -1, 0, -1, -1 };
|
|
|
|
VkVideoEncodeAV1PredictionModeKHR pred_mode;
|
|
VkVideoEncodeAV1RateControlGroupKHR rc_group;
|
|
int lr_unit_shift = 0;
|
|
int lr_uv_shift = 0;
|
|
|
|
ap->ext_header = (StdVideoEncodeAV1ExtensionHeader) {
|
|
.temporal_id = 0,
|
|
.spatial_id = 0,
|
|
};
|
|
|
|
*tile_info = (StdVideoAV1TileInfo) {
|
|
.flags = (StdVideoAV1TileInfoFlags) {
|
|
.uniform_tile_spacing_flag = enc->uniform_tile,
|
|
},
|
|
.TileCols = enc->tile_cols,
|
|
.TileRows = enc->tile_rows,
|
|
.context_update_tile_id = 0,
|
|
.tile_size_bytes_minus_1 = 0,
|
|
};
|
|
|
|
for (int i = 0; i < STD_VIDEO_AV1_TOTAL_REFS_PER_FRAME; i++) {
|
|
global_motion->GmType[i] = 0;
|
|
for (int j = 0; j < STD_VIDEO_AV1_GLOBAL_MOTION_PARAMS; j++) {
|
|
global_motion->gm_params[i][j] = 0;
|
|
}
|
|
}
|
|
|
|
for (int i = 0; i < STD_VIDEO_AV1_REFS_PER_FRAME; i++)
|
|
name_slots[i] = -1;
|
|
|
|
*loop_restoration = (StdVideoAV1LoopRestoration) {
|
|
.FrameRestorationType[0] = STD_VIDEO_AV1_FRAME_RESTORATION_TYPE_NONE,
|
|
.FrameRestorationType[1] = STD_VIDEO_AV1_FRAME_RESTORATION_TYPE_NONE,
|
|
.FrameRestorationType[2] = STD_VIDEO_AV1_FRAME_RESTORATION_TYPE_NONE,
|
|
.LoopRestorationSize[0] = 1 + lr_unit_shift,
|
|
.LoopRestorationSize[1] = 1 + lr_unit_shift - lr_uv_shift,
|
|
.LoopRestorationSize[2] = 1 + lr_unit_shift - lr_uv_shift,
|
|
};
|
|
|
|
*cdef = (StdVideoAV1CDEF) {
|
|
.cdef_damping_minus_3 = 0,
|
|
.cdef_bits = 0,
|
|
};
|
|
|
|
for (int i = 0; i < STD_VIDEO_AV1_MAX_SEGMENTS; i++) {
|
|
segmentation->FeatureEnabled[i] = 0x0;
|
|
for (int j = 0; j < STD_VIDEO_AV1_SEG_LVL_MAX; j++) {
|
|
segmentation->FeatureEnabled[i] |= 0x0;
|
|
segmentation->FeatureData[i][j] = 0;
|
|
}
|
|
}
|
|
|
|
*loop_filter = (StdVideoAV1LoopFilter) {
|
|
.flags = (StdVideoAV1LoopFilterFlags) {
|
|
.loop_filter_delta_enabled = 0,
|
|
.loop_filter_delta_update = 0,
|
|
},
|
|
.loop_filter_level = { 0 },
|
|
.loop_filter_sharpness = 0,
|
|
.update_ref_delta = 0,
|
|
.loop_filter_ref_deltas = { 0 },
|
|
.update_mode_delta = 0,
|
|
.loop_filter_mode_deltas = { 0 },
|
|
};
|
|
loop_filter->update_mode_delta = 1;
|
|
memcpy(loop_filter->loop_filter_ref_deltas, default_loop_filter_ref_deltas,
|
|
STD_VIDEO_AV1_TOTAL_REFS_PER_FRAME * sizeof(int8_t));
|
|
|
|
*quantization = (StdVideoAV1Quantization) {
|
|
.flags = (StdVideoAV1QuantizationFlags) {
|
|
.using_qmatrix = 0,
|
|
.diff_uv_delta = 0,
|
|
/* Reserved */
|
|
},
|
|
.base_q_idx = 0, /* Set later */
|
|
.DeltaQYDc = 0,
|
|
.DeltaQUDc = 0,
|
|
.DeltaQUAc = 0,
|
|
.DeltaQVDc = 0,
|
|
.DeltaQVAc = 0,
|
|
.qm_y = 0,
|
|
.qm_u = 0,
|
|
.qm_v = 0,
|
|
};
|
|
|
|
ref_slot = (VkVideoReferenceSlotInfoKHR *)encode_info->pSetupReferenceSlot;
|
|
ap->av1pic_info = (StdVideoEncodeAV1PictureInfo) {
|
|
.flags = (StdVideoEncodeAV1PictureInfoFlags) {
|
|
.error_resilient_mode = (pic->type == FF_HW_PICTURE_TYPE_I ||
|
|
pic->type == FF_HW_PICTURE_TYPE_IDR) &&
|
|
(pic->display_order <= pic->encode_order),
|
|
.disable_cdf_update = 0,
|
|
.use_superres = 0,
|
|
.render_and_frame_size_different = 0,
|
|
.allow_screen_content_tools = 0,
|
|
.is_filter_switchable = 0,
|
|
.force_integer_mv = 0,
|
|
.frame_size_override_flag = 0,
|
|
.buffer_removal_time_present_flag = 0,
|
|
.allow_intrabc = 0,
|
|
.frame_refs_short_signaling = 0,
|
|
.allow_high_precision_mv = 0,
|
|
.is_motion_mode_switchable = 0,
|
|
.use_ref_frame_mvs = 0,
|
|
.disable_frame_end_update_cdf = 0,
|
|
.allow_warped_motion = 0,
|
|
.reduced_tx_set = 0,
|
|
.skip_mode_present = 0,
|
|
.delta_q_present = 0,
|
|
.delta_lf_present = 0,
|
|
.delta_lf_multi = 0,
|
|
.segmentation_enabled = 0,
|
|
.segmentation_update_map = 0,
|
|
.segmentation_temporal_update = 0,
|
|
.segmentation_update_data = 0,
|
|
.UsesLr = 0,
|
|
.usesChromaLr = 0,
|
|
.show_frame = pic->display_order <= pic->encode_order,
|
|
.showable_frame = 0,
|
|
/* Reserved */
|
|
},
|
|
.frame_type = 0, // set later
|
|
.frame_presentation_time = 0,
|
|
.current_frame_id = ref_slot->slotIndex,
|
|
.order_hint = 0, // set later
|
|
.primary_ref_frame = 0, // set later
|
|
.refresh_frame_flags = 0x0, // set later
|
|
.coded_denom = 0,
|
|
.render_width_minus_1 = base_ctx->surface_width - 1,
|
|
.render_height_minus_1 = base_ctx->surface_height - 1,
|
|
.interpolation_filter = 0,
|
|
.TxMode = STD_VIDEO_AV1_TX_MODE_SELECT,
|
|
.delta_q_res = 0,
|
|
.delta_lf_res = 0,
|
|
.ref_order_hint = { 0 }, // set later
|
|
.ref_frame_idx = { 0 }, // set later
|
|
/* Reserved */
|
|
.delta_frame_id_minus_1 = { 0 },
|
|
|
|
// .pTileInfo = tile_info, TODO FIX
|
|
.pQuantization = quantization,
|
|
.pSegmentation = segmentation,
|
|
.pLoopFilter = loop_filter,
|
|
.pCDEF = cdef,
|
|
.pLoopRestoration = loop_restoration,
|
|
.pGlobalMotion = global_motion,
|
|
.pExtensionHeader = &ap->ext_header,
|
|
.pBufferRemovalTimes = NULL,
|
|
};
|
|
|
|
switch (pic->type) {
|
|
case FF_HW_PICTURE_TYPE_I:
|
|
case FF_HW_PICTURE_TYPE_IDR:
|
|
av_assert0(pic->nb_refs[0] == 0 || pic->nb_refs[1]);
|
|
ap->av1pic_info.frame_type = STD_VIDEO_AV1_FRAME_TYPE_KEY;
|
|
ap->av1pic_info.refresh_frame_flags = 0xFF;
|
|
quantization->base_q_idx = enc->q_idx_idr;
|
|
ap->slot = 0;
|
|
ap->last_idr_frame = pic->display_order;
|
|
pred_mode = VK_VIDEO_ENCODE_AV1_PREDICTION_MODE_INTRA_ONLY_KHR;
|
|
rc_group = VK_VIDEO_ENCODE_AV1_RATE_CONTROL_GROUP_INTRA_KHR;
|
|
break;
|
|
case FF_HW_PICTURE_TYPE_P:
|
|
ref = pic->refs[0][pic->nb_refs[0] - 1];
|
|
ap_ref = ref->codec_priv;
|
|
|
|
ap->av1pic_info.frame_type = STD_VIDEO_AV1_FRAME_TYPE_INTER;
|
|
quantization->base_q_idx = enc->q_idx_p;
|
|
|
|
ap->last_idr_frame = ap_ref->last_idr_frame;
|
|
ap->slot = !ap_ref->slot;
|
|
|
|
ap->av1pic_info.refresh_frame_flags = 1 << ap->slot;
|
|
|
|
/** set the nearest frame in L0 as all reference frame. */
|
|
for (int i = 0; i < AV1_REFS_PER_FRAME; i++)
|
|
ap->av1pic_info.ref_frame_idx[i] = ap_ref->slot;
|
|
|
|
ap->av1pic_info.primary_ref_frame = ap_ref->slot;
|
|
ap->av1pic_info.ref_order_hint[ap_ref->slot] = ref->display_order - ap_ref->last_idr_frame;
|
|
rc_group = VK_VIDEO_ENCODE_AV1_RATE_CONTROL_GROUP_PREDICTIVE_KHR;
|
|
pred_mode = VK_VIDEO_ENCODE_AV1_PREDICTION_MODE_SINGLE_REFERENCE_KHR;
|
|
ref_name_mask = enc->caps.singleReferenceNameMask;
|
|
set_name_slot(ap_ref->av1pic_info.current_frame_id, name_slots, ref_name_mask, 0);
|
|
|
|
// vpic->ref_frame_ctrl_l0.fields.search_idx0 = AV1_REF_FRAME_LAST;
|
|
|
|
/** set the 2nd nearest frame in L0 as Golden frame. */
|
|
if ((pic->nb_refs[0] > 1) &&
|
|
((enc->caps.maxSingleReferenceCount > 1) ||
|
|
(enc->caps.maxUnidirectionalCompoundReferenceCount > 0))) {
|
|
if (enc->caps.maxUnidirectionalCompoundReferenceCount) {
|
|
pred_mode = VK_VIDEO_ENCODE_AV1_PREDICTION_MODE_UNIDIRECTIONAL_COMPOUND_KHR;
|
|
ref_name_mask = enc->caps.unidirectionalCompoundReferenceNameMask;
|
|
}
|
|
ref = pic->refs[0][pic->nb_refs[0] - 2];
|
|
ap_ref = ref->codec_priv;
|
|
ap->av1pic_info.ref_frame_idx[3] = ap_ref->slot;
|
|
ap->av1pic_info.ref_order_hint[ap_ref->slot] = ref->display_order - ap_ref->last_idr_frame;
|
|
// vpic->ref_frame_ctrl_l0.fields.search_idx1 = AV1_REF_FRAME_GOLDEN;
|
|
set_name_slot(ap_ref->av1pic_info.current_frame_id, name_slots, ref_name_mask, 0);
|
|
}
|
|
break;
|
|
case FF_HW_PICTURE_TYPE_B:
|
|
ap->av1pic_info.frame_type = STD_VIDEO_AV1_FRAME_TYPE_INTER;
|
|
quantization->base_q_idx = enc->q_idx_b;
|
|
ap->av1pic_info.refresh_frame_flags = 0x0;
|
|
|
|
rc_group = VK_VIDEO_ENCODE_AV1_RATE_CONTROL_GROUP_BIPREDICTIVE_KHR;
|
|
pred_mode = VK_VIDEO_ENCODE_AV1_PREDICTION_MODE_BIDIRECTIONAL_COMPOUND_KHR;
|
|
ref_name_mask = enc->caps.bidirectionalCompoundReferenceNameMask;
|
|
|
|
// fh->reference_select = 1;
|
|
/** B frame will not be referenced, disable its recon frame. */
|
|
// vpic->picture_flags.bits.disable_frame_recon = 1;
|
|
|
|
/** Use LAST_FRAME and BWDREF_FRAME for reference. */
|
|
// vpic->ref_frame_ctrl_l0.fields.search_idx0 = AV1_REF_FRAME_LAST;
|
|
// vpic->ref_frame_ctrl_l1.fields.search_idx0 = AV1_REF_FRAME_BWDREF;
|
|
|
|
ref = pic->refs[0][pic->nb_refs[0] - 1];
|
|
ap_ref = ref->codec_priv;
|
|
ap->last_idr_frame = ap_ref->last_idr_frame;
|
|
ap->av1pic_info.primary_ref_frame = ap_ref->slot;
|
|
ap->av1pic_info.ref_order_hint[ap_ref->slot] = ref->display_order - ap_ref->last_idr_frame;
|
|
for (int i = 0; i < AV1_REF_FRAME_GOLDEN; i++)
|
|
ap->av1pic_info.ref_frame_idx[i] = ap_ref->slot;
|
|
set_name_slot(ap_ref->av1pic_info.current_frame_id, name_slots, ref_name_mask, 0);
|
|
|
|
ref = pic->refs[1][pic->nb_refs[1] - 1];
|
|
ap_ref = ref->codec_priv;
|
|
ap->av1pic_info.ref_order_hint[ap_ref->slot] = ref->display_order - ap_ref->last_idr_frame;
|
|
for (int i = AV1_REF_FRAME_GOLDEN; i < AV1_REFS_PER_FRAME; i++)
|
|
ap->av1pic_info.ref_frame_idx[i] = ap_ref->slot;
|
|
set_name_slot(ap_ref->av1pic_info.current_frame_id, name_slots, ref_name_mask, 1);
|
|
break;
|
|
}
|
|
|
|
ap->av1pic_info.flags.showable_frame = ap->av1pic_info.frame_type != STD_VIDEO_AV1_FRAME_TYPE_KEY;
|
|
ap->av1pic_info.order_hint = pic->display_order - ap->last_idr_frame;
|
|
|
|
ap->vkav1pic_info = (VkVideoEncodeAV1PictureInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_PICTURE_INFO_KHR,
|
|
.pNext = NULL,
|
|
.predictionMode = pred_mode,
|
|
.rateControlGroup = rc_group,
|
|
.constantQIndex = quantization->base_q_idx,
|
|
.pStdPictureInfo = &ap->av1pic_info,
|
|
// .referenceNameSlotIndices is set below
|
|
.primaryReferenceCdfOnly = 0,
|
|
.generateObuExtensionHeader = 0,
|
|
};
|
|
encode_info->pNext = &ap->vkav1pic_info;
|
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(ap->vkav1pic_info.referenceNameSlotIndices); i++)
|
|
ap->vkav1pic_info.referenceNameSlotIndices[i] = name_slots[i];
|
|
|
|
ref_slot = (VkVideoReferenceSlotInfoKHR *)encode_info->pSetupReferenceSlot;
|
|
ref_slot->pNext = &ap->vkav1dpb_info;
|
|
|
|
ap->av1dpb_info = (StdVideoEncodeAV1ReferenceInfo) {
|
|
.flags = (StdVideoEncodeAV1ReferenceInfoFlags) {
|
|
.disable_frame_end_update_cdf = 0,
|
|
.segmentation_enabled = 0,
|
|
/* Reserved */
|
|
},
|
|
.RefFrameId = ref_slot->slotIndex,
|
|
.frame_type = ap->av1pic_info.frame_type,
|
|
.OrderHint = pic->display_order - ap->last_idr_frame,
|
|
/* Reserved */
|
|
.pExtensionHeader = &ap->ext_header,
|
|
};
|
|
|
|
ap->vkav1dpb_info = (VkVideoEncodeAV1DpbSlotInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_DPB_SLOT_INFO_KHR,
|
|
.pStdReferenceInfo = &ap->av1dpb_info,
|
|
};
|
|
|
|
ap->units_needed = 0;
|
|
if (pic->type == FF_HW_PICTURE_TYPE_IDR) {
|
|
AVFrameSideData *sd = NULL;
|
|
if (enc->unit_elems & UNIT_MASTERING_DISPLAY)
|
|
sd = av_frame_get_side_data(pic->input_image,
|
|
AV_FRAME_DATA_MASTERING_DISPLAY_METADATA);
|
|
if (sd) {
|
|
AVMasteringDisplayMetadata *mdm =
|
|
(AVMasteringDisplayMetadata *)sd->data;
|
|
if (mdm->has_primaries && mdm->has_luminance) {
|
|
AV1RawOBU *obu = &enc->meta_mastering_obu;
|
|
AV1RawMetadata *md = &obu->obu.metadata;
|
|
AV1RawMetadataHDRMDCV *mdcv = &md->metadata.hdr_mdcv;
|
|
const int chroma_den = 1 << 16;
|
|
const int max_luma_den = 1 << 8;
|
|
const int min_luma_den = 1 << 14;
|
|
|
|
memset(obu, 0, sizeof(*obu));
|
|
obu->header.obu_type = AV1_OBU_METADATA;
|
|
md->metadata_type = AV1_METADATA_TYPE_HDR_MDCV;
|
|
|
|
for (int i = 0; i < 3; i++) {
|
|
mdcv->primary_chromaticity_x[i] =
|
|
av_rescale(mdm->display_primaries[i][0].num, chroma_den,
|
|
mdm->display_primaries[i][0].den);
|
|
mdcv->primary_chromaticity_y[i] =
|
|
av_rescale(mdm->display_primaries[i][1].num, chroma_den,
|
|
mdm->display_primaries[i][1].den);
|
|
}
|
|
|
|
mdcv->white_point_chromaticity_x =
|
|
av_rescale(mdm->white_point[0].num, chroma_den,
|
|
mdm->white_point[0].den);
|
|
mdcv->white_point_chromaticity_y =
|
|
av_rescale(mdm->white_point[1].num, chroma_den,
|
|
mdm->white_point[1].den);
|
|
|
|
mdcv->luminance_max =
|
|
av_rescale(mdm->max_luminance.num, max_luma_den,
|
|
mdm->max_luminance.den);
|
|
mdcv->luminance_min =
|
|
av_rescale(mdm->min_luminance.num, min_luma_den,
|
|
mdm->min_luminance.den);
|
|
ap->units_needed |= UNIT_MASTERING_DISPLAY;
|
|
}
|
|
}
|
|
|
|
if (enc->unit_elems & UNIT_CONTENT_LIGHT_LEVEL)
|
|
sd = av_frame_get_side_data(pic->input_image,
|
|
AV_FRAME_DATA_CONTENT_LIGHT_LEVEL);
|
|
if (sd) {
|
|
AVContentLightMetadata *cllm = (AVContentLightMetadata *)sd->data;
|
|
AV1RawOBU *obu = &enc->meta_cll_obu;
|
|
AV1RawMetadata *md = &obu->obu.metadata;
|
|
AV1RawMetadataHDRCLL *cll = &md->metadata.hdr_cll;
|
|
|
|
memset(obu, 0, sizeof(*obu));
|
|
obu->header.obu_type = AV1_OBU_METADATA;
|
|
md->metadata_type = AV1_METADATA_TYPE_HDR_CLL;
|
|
cll->max_cll = cllm->MaxCLL;
|
|
cll->max_fall = cllm->MaxFALL;
|
|
|
|
ap->units_needed |= UNIT_CONTENT_LIGHT_LEVEL;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int init_profile(AVCodecContext *avctx,
|
|
VkVideoProfileInfoKHR *profile, void *pnext)
|
|
{
|
|
VkResult ret;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFVulkanContext *s = &ctx->s;
|
|
FFVulkanFunctions *vk = &ctx->s.vkfn;
|
|
FFHWBaseEncodeContext *base_ctx = &ctx->base;
|
|
|
|
VkVideoEncodeAV1CapabilitiesKHR av1_caps = {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_CAPABILITIES_KHR,
|
|
};
|
|
VkVideoEncodeCapabilitiesKHR enc_caps = {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_CAPABILITIES_KHR,
|
|
.pNext = &av1_caps,
|
|
};
|
|
VkVideoCapabilitiesKHR caps = {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_CAPABILITIES_KHR,
|
|
.pNext = &enc_caps,
|
|
};
|
|
|
|
/* In order of preference */
|
|
int last_supported = AV_PROFILE_UNKNOWN;
|
|
static const int known_profiles[] = {
|
|
AV_PROFILE_AV1_MAIN,
|
|
AV_PROFILE_AV1_HIGH,
|
|
AV_PROFILE_AV1_PROFESSIONAL,
|
|
};
|
|
int nb_profiles = FF_ARRAY_ELEMS(known_profiles);
|
|
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->frames->sw_format);
|
|
if (!desc)
|
|
return AVERROR(EINVAL);
|
|
|
|
if (s->frames->sw_format == AV_PIX_FMT_NV12 ||
|
|
s->frames->sw_format == AV_PIX_FMT_P010)
|
|
nb_profiles = 1;
|
|
|
|
enc->profile = (VkVideoEncodeAV1ProfileInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_PROFILE_INFO_KHR,
|
|
.pNext = pnext,
|
|
.stdProfile = ff_vk_av1_profile_to_vk(avctx->profile),
|
|
};
|
|
profile->pNext = &enc->profile;
|
|
|
|
/* Set level */
|
|
if (avctx->level == AV_LEVEL_UNKNOWN) {
|
|
const AV1LevelDescriptor *level;
|
|
float framerate = 0.0;
|
|
|
|
if (avctx->framerate.num > 0 && avctx->framerate.den > 0)
|
|
framerate = av_q2d(avctx->framerate);
|
|
|
|
level = ff_av1_guess_level(avctx->bit_rate, enc->seq_tier,
|
|
base_ctx->surface_width, base_ctx->surface_height,
|
|
enc->tile_rows * enc->tile_cols,
|
|
enc->tile_cols, framerate);
|
|
if (level) {
|
|
av_log(avctx, AV_LOG_VERBOSE, "Using level %s.\n", level->name);
|
|
enc->seq_level_idx = ff_vk_av1_level_to_vk(level->level_idx);
|
|
} else {
|
|
av_log(avctx, AV_LOG_VERBOSE, "Stream will not conform to "
|
|
"any normal level, using level 7.3 by default.\n");
|
|
enc->seq_level_idx = STD_VIDEO_AV1_LEVEL_7_3;
|
|
enc->seq_tier = 1;
|
|
}
|
|
} else {
|
|
enc->seq_level_idx = ff_vk_av1_level_to_vk(avctx->level);
|
|
}
|
|
|
|
/* User has explicitly specified a profile. */
|
|
if (avctx->profile != AV_PROFILE_UNKNOWN)
|
|
return 0;
|
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Supported profiles:\n");
|
|
for (int i = 0; i < nb_profiles; i++) {
|
|
enc->profile.stdProfile = ff_vk_av1_profile_to_vk(known_profiles[i]);
|
|
ret = vk->GetPhysicalDeviceVideoCapabilitiesKHR(s->hwctx->phys_dev,
|
|
profile,
|
|
&caps);
|
|
if (ret == VK_SUCCESS) {
|
|
av_log(avctx, AV_LOG_DEBUG, " %s\n",
|
|
avcodec_profile_name(avctx->codec_id, known_profiles[i]));
|
|
last_supported = known_profiles[i];
|
|
}
|
|
}
|
|
|
|
if (last_supported == AV_PROFILE_UNKNOWN) {
|
|
av_log(avctx, AV_LOG_ERROR, "No supported profiles for given format\n");
|
|
return AVERROR(ENOTSUP);
|
|
}
|
|
|
|
enc->profile.stdProfile = ff_vk_av1_profile_to_vk(last_supported);
|
|
av_log(avctx, AV_LOG_VERBOSE, "Using profile %s\n",
|
|
avcodec_profile_name(avctx->codec_id, last_supported));
|
|
avctx->profile = last_supported;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int init_enc_options(AVCodecContext *avctx)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
|
|
if (avctx->rc_buffer_size)
|
|
enc->hrd_buffer_size = avctx->rc_buffer_size;
|
|
else if (avctx->rc_max_rate > 0)
|
|
enc->hrd_buffer_size = avctx->rc_max_rate;
|
|
else
|
|
enc->hrd_buffer_size = avctx->bit_rate;
|
|
|
|
if (avctx->rc_initial_buffer_occupancy) {
|
|
if (avctx->rc_initial_buffer_occupancy > enc->hrd_buffer_size) {
|
|
av_log(avctx, AV_LOG_ERROR, "Invalid RC buffer settings: "
|
|
"must have initial buffer size (%d) <= "
|
|
"buffer size (%"PRId64").\n",
|
|
avctx->rc_initial_buffer_occupancy, enc->hrd_buffer_size);
|
|
return AVERROR(EINVAL);
|
|
}
|
|
enc->initial_buffer_fullness = avctx->rc_initial_buffer_occupancy;
|
|
} else {
|
|
enc->initial_buffer_fullness = enc->hrd_buffer_size * 3 / 4;
|
|
}
|
|
|
|
if (enc->common.opts.rc_mode == VK_VIDEO_ENCODE_RATE_CONTROL_MODE_DISABLED_BIT_KHR) {
|
|
enc->q_idx_p = av_clip(enc->common.opts.qp,
|
|
enc->caps.minQIndex, enc->caps.maxQIndex);
|
|
if (fabs(avctx->i_quant_factor) > 0.0)
|
|
enc->q_idx_idr =
|
|
av_clip((fabs(avctx->i_quant_factor) * enc->q_idx_p +
|
|
avctx->i_quant_offset) + 0.5,
|
|
0, 255);
|
|
else
|
|
enc->q_idx_idr = enc->q_idx_p;
|
|
|
|
if (fabs(avctx->b_quant_factor) > 0.0)
|
|
enc->q_idx_b =
|
|
av_clip((fabs(avctx->b_quant_factor) * enc->q_idx_p +
|
|
avctx->b_quant_offset) + 0.5,
|
|
0, 255);
|
|
else
|
|
enc->q_idx_b = enc->q_idx_p;
|
|
} else {
|
|
/** Arbitrary value */
|
|
enc->q_idx_idr = enc->q_idx_p = enc->q_idx_b = 128;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int init_sequence_headers(AVCodecContext *avctx)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFVulkanContext *s = &ctx->s;
|
|
FFHWBaseEncodeContext *base_ctx = &ctx->base;
|
|
|
|
AV1RawOBU *seq_obu = &enc->seq_hdr_obu;
|
|
AV1RawSequenceHeader *seq = &seq_obu->obu.sequence_header;
|
|
|
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(s->frames->sw_format);
|
|
if (!desc)
|
|
return AVERROR(EINVAL);
|
|
|
|
seq_obu->header.obu_type = AV1_OBU_SEQUENCE_HEADER;
|
|
*seq = (AV1RawSequenceHeader) {
|
|
.seq_profile = avctx->profile,
|
|
.seq_force_integer_mv = seq->seq_force_screen_content_tools ?
|
|
AV1_SELECT_SCREEN_CONTENT_TOOLS :
|
|
AV1_SELECT_INTEGER_MV,
|
|
.frame_width_bits_minus_1 = av_log2(base_ctx->surface_width),
|
|
.frame_height_bits_minus_1 = av_log2(base_ctx->surface_height),
|
|
.max_frame_width_minus_1 = base_ctx->surface_width - 1,
|
|
.max_frame_height_minus_1 = base_ctx->surface_height - 1,
|
|
.enable_order_hint = 1,
|
|
.order_hint_bits_minus_1 = av_clip_intp2(av_log2(ctx->base.gop_size), 3),
|
|
.use_128x128_superblock = !!(enc->caps.superblockSizes & VK_VIDEO_ENCODE_AV1_SUPERBLOCK_SIZE_128_BIT_KHR),
|
|
.color_config = (AV1RawColorConfig) {
|
|
.high_bitdepth = desc->comp[0].depth > 8,
|
|
.color_primaries = avctx->color_primaries,
|
|
.transfer_characteristics = avctx->color_trc,
|
|
.matrix_coefficients = avctx->colorspace,
|
|
.color_description_present_flag = (avctx->color_primaries != AVCOL_PRI_UNSPECIFIED ||
|
|
avctx->color_trc != AVCOL_TRC_UNSPECIFIED ||
|
|
avctx->colorspace != AVCOL_SPC_UNSPECIFIED),
|
|
.subsampling_x = desc->log2_chroma_w,
|
|
.subsampling_y = desc->log2_chroma_h,
|
|
.chroma_sample_position = avctx->chroma_sample_location == AVCHROMA_LOC_LEFT ?
|
|
AV1_CSP_VERTICAL :
|
|
avctx->chroma_sample_location == AVCHROMA_LOC_TOPLEFT ?
|
|
AV1_CSP_COLOCATED :
|
|
AV1_CSP_UNKNOWN,
|
|
},
|
|
|
|
/* Operating point */
|
|
.seq_tier = { enc->seq_tier },
|
|
.seq_level_idx = { enc->seq_level_idx },
|
|
.decoder_buffer_delay = { base_ctx->decode_delay },
|
|
.encoder_buffer_delay = { base_ctx->output_delay },
|
|
.operating_points_cnt_minus_1 = 1 - 1,
|
|
};
|
|
|
|
return 0;
|
|
}
|
|
|
|
typedef struct VulkanAV1Units {
|
|
StdVideoAV1SequenceHeader seq_hdr;
|
|
StdVideoAV1TimingInfo timing_info;
|
|
StdVideoAV1ColorConfig color_config;
|
|
|
|
StdVideoEncodeAV1DecoderModelInfo decoder_model;
|
|
StdVideoEncodeAV1OperatingPointInfo operating_points[AV1_MAX_OPERATING_POINTS];
|
|
int nb_operating_points;
|
|
} VulkanAV1Units;
|
|
|
|
static av_cold int base_unit_to_vk(AVCodecContext *avctx, VulkanAV1Units *vk_units)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
|
|
AV1RawOBU *seq_obu = &enc->seq_hdr_obu;
|
|
AV1RawSequenceHeader *seq = &seq_obu->obu.sequence_header;
|
|
|
|
StdVideoAV1SequenceHeader *seq_hdr = &vk_units->seq_hdr;
|
|
StdVideoAV1TimingInfo *timing_info = &vk_units->timing_info;
|
|
StdVideoAV1ColorConfig *color_config = &vk_units->color_config;
|
|
|
|
StdVideoEncodeAV1OperatingPointInfo *operating_points = vk_units->operating_points;
|
|
|
|
*timing_info = (StdVideoAV1TimingInfo) {
|
|
.flags = (StdVideoAV1TimingInfoFlags) {
|
|
.equal_picture_interval = seq->timing_info.equal_picture_interval,
|
|
},
|
|
.num_units_in_display_tick = seq->timing_info.num_units_in_display_tick,
|
|
.time_scale = seq->timing_info.time_scale,
|
|
.num_ticks_per_picture_minus_1 = seq->timing_info.num_ticks_per_picture_minus_1,
|
|
};
|
|
|
|
*color_config = (StdVideoAV1ColorConfig) {
|
|
.flags = (StdVideoAV1ColorConfigFlags) {
|
|
.mono_chrome = seq->color_config.mono_chrome,
|
|
.color_range = seq->color_config.color_range,
|
|
.separate_uv_delta_q = seq->color_config.separate_uv_delta_q,
|
|
},
|
|
.BitDepth = seq->color_config.twelve_bit ? 12 :
|
|
seq->color_config.high_bitdepth ? 10 : 8,
|
|
.subsampling_x = seq->color_config.subsampling_x,
|
|
.subsampling_y = seq->color_config.subsampling_y,
|
|
.color_primaries = seq->color_config.color_primaries,
|
|
.transfer_characteristics = seq->color_config.transfer_characteristics,
|
|
.matrix_coefficients = seq->color_config.matrix_coefficients,
|
|
};
|
|
|
|
*seq_hdr = (StdVideoAV1SequenceHeader) {
|
|
.flags = (StdVideoAV1SequenceHeaderFlags) {
|
|
.still_picture = seq->still_picture,
|
|
.reduced_still_picture_header = seq->reduced_still_picture_header,
|
|
.use_128x128_superblock = seq->use_128x128_superblock,
|
|
.enable_filter_intra = seq->enable_filter_intra,
|
|
.enable_intra_edge_filter = seq->enable_intra_edge_filter,
|
|
.enable_interintra_compound = seq->enable_interintra_compound,
|
|
.enable_masked_compound = seq->enable_masked_compound,
|
|
.enable_warped_motion = seq->enable_warped_motion,
|
|
.enable_dual_filter = seq->enable_dual_filter,
|
|
.enable_order_hint = seq->enable_order_hint,
|
|
.enable_jnt_comp = seq->enable_jnt_comp,
|
|
.enable_ref_frame_mvs = seq->enable_ref_frame_mvs,
|
|
.frame_id_numbers_present_flag = seq->frame_id_numbers_present_flag,
|
|
.enable_superres = seq->enable_superres,
|
|
.enable_cdef = seq->enable_cdef,
|
|
.enable_restoration = seq->enable_restoration,
|
|
.film_grain_params_present = seq->film_grain_params_present,
|
|
.timing_info_present_flag = seq->timing_info_present_flag,
|
|
.initial_display_delay_present_flag = seq->initial_display_delay_present_flag,
|
|
},
|
|
.seq_profile = seq->seq_profile,
|
|
.frame_width_bits_minus_1 = seq->frame_width_bits_minus_1,
|
|
.frame_height_bits_minus_1 = seq->frame_height_bits_minus_1,
|
|
.max_frame_width_minus_1 = seq->max_frame_width_minus_1,
|
|
.max_frame_height_minus_1 = seq->max_frame_height_minus_1,
|
|
.delta_frame_id_length_minus_2 = seq->delta_frame_id_length_minus_2,
|
|
.additional_frame_id_length_minus_1 = seq->additional_frame_id_length_minus_1,
|
|
.order_hint_bits_minus_1 = seq->order_hint_bits_minus_1,
|
|
.seq_force_integer_mv = seq->seq_force_integer_mv,
|
|
.seq_force_screen_content_tools = seq->seq_force_screen_content_tools,
|
|
.pTimingInfo = timing_info,
|
|
.pColorConfig = color_config,
|
|
};
|
|
|
|
for (int i = 0; i <= seq->operating_points_cnt_minus_1; i++) {
|
|
operating_points[i] = (StdVideoEncodeAV1OperatingPointInfo) {
|
|
.flags = (StdVideoEncodeAV1OperatingPointInfoFlags) {
|
|
.decoder_model_present_for_this_op = seq->decoder_model_present_for_this_op[i],
|
|
.low_delay_mode_flag = seq->low_delay_mode_flag[i],
|
|
.initial_display_delay_present_for_this_op = seq->initial_display_delay_present_for_this_op[i],
|
|
/* Reserved */
|
|
},
|
|
.operating_point_idc = seq->operating_point_idc[i],
|
|
.seq_level_idx = seq->seq_level_idx[i],
|
|
.seq_tier = seq->seq_tier[i],
|
|
.decoder_buffer_delay = seq->decoder_buffer_delay[i],
|
|
.encoder_buffer_delay = seq->encoder_buffer_delay[i],
|
|
.initial_display_delay_minus_1 = seq->initial_display_delay_minus_1[i],
|
|
};
|
|
}
|
|
vk_units->nb_operating_points = seq->operating_points_cnt_minus_1 + 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int create_session_params(AVCodecContext *avctx)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFVulkanContext *s = &ctx->s;
|
|
FFVulkanFunctions *vk = &ctx->s.vkfn;
|
|
|
|
VulkanAV1Units vk_units = { 0 };
|
|
|
|
VkVideoEncodeAV1SessionParametersCreateInfoKHR av1_params;
|
|
|
|
/* Convert it to Vulkan */
|
|
err = base_unit_to_vk(avctx, &vk_units);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Unable to convert sequence header to Vulkan: %s\n",
|
|
av_err2str(err));
|
|
return err;
|
|
}
|
|
|
|
/* Destroy the session params */
|
|
if (ctx->session_params)
|
|
vk->DestroyVideoSessionParametersKHR(s->hwctx->act_dev,
|
|
ctx->session_params,
|
|
s->hwctx->alloc);
|
|
|
|
av1_params = (VkVideoEncodeAV1SessionParametersCreateInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_SESSION_PARAMETERS_CREATE_INFO_KHR,
|
|
.pStdSequenceHeader = &vk_units.seq_hdr,
|
|
.pStdDecoderModelInfo = &vk_units.decoder_model,
|
|
.pStdOperatingPoints = vk_units.operating_points,
|
|
.stdOperatingPointCount = vk_units.nb_operating_points,
|
|
};
|
|
|
|
return ff_vulkan_encode_create_session_params(avctx, ctx, &av1_params);
|
|
}
|
|
|
|
static int parse_feedback_units(AVCodecContext *avctx,
|
|
const uint8_t *data, size_t size)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
AV1RawOBU *seq_obu = &enc->seq_hdr_obu;
|
|
AV1RawSequenceHeader *seq = &seq_obu->obu.sequence_header;
|
|
|
|
CodedBitstreamContext *cbs;
|
|
CodedBitstreamFragment obu = { 0 };
|
|
|
|
err = ff_cbs_init(&cbs, AV_CODEC_ID_AV1, avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
err = ff_cbs_read(cbs, &obu, NULL, data, size);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Unable to parse feedback units, bad drivers: %s\n",
|
|
av_err2str(err));
|
|
return err;
|
|
}
|
|
|
|
/* If PPS has an override, just copy it entirely. */
|
|
for (int i = 0; i < obu.nb_units; i++) {
|
|
if (obu.units[i].type == AV1_OBU_SEQUENCE_HEADER) {
|
|
AV1RawOBU *f_seq_obu = obu.units[i].content;
|
|
AV1RawSequenceHeader *f_seq = &f_seq_obu->obu.sequence_header;
|
|
seq->frame_width_bits_minus_1 = f_seq->frame_width_bits_minus_1;
|
|
seq->frame_height_bits_minus_1 = f_seq->frame_height_bits_minus_1;
|
|
seq->max_frame_width_minus_1 = f_seq->max_frame_width_minus_1;
|
|
seq->max_frame_height_minus_1 = f_seq->max_frame_height_minus_1;
|
|
seq->seq_choose_screen_content_tools = f_seq->seq_choose_screen_content_tools;
|
|
seq->seq_force_screen_content_tools = f_seq->seq_force_screen_content_tools;
|
|
seq->seq_choose_integer_mv = f_seq->seq_choose_integer_mv;
|
|
seq->seq_force_integer_mv = f_seq->seq_force_integer_mv;
|
|
}
|
|
}
|
|
|
|
ff_cbs_fragment_free(&obu);
|
|
ff_cbs_close(&cbs);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int init_base_units(AVCodecContext *avctx)
|
|
{
|
|
int err;
|
|
VkResult ret;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFVulkanContext *s = &ctx->s;
|
|
FFVulkanFunctions *vk = &ctx->s.vkfn;
|
|
|
|
VkVideoEncodeSessionParametersGetInfoKHR params_info;
|
|
VkVideoEncodeSessionParametersFeedbackInfoKHR params_feedback;
|
|
|
|
void *data = NULL;
|
|
size_t data_size = 0;
|
|
|
|
/* Generate SPS/PPS unit info */
|
|
err = init_sequence_headers(avctx);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Unable to initialize sequence header: %s\n",
|
|
av_err2str(err));
|
|
return err;
|
|
}
|
|
|
|
/* Create session parameters from them */
|
|
err = create_session_params(avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
params_info = (VkVideoEncodeSessionParametersGetInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_SESSION_PARAMETERS_GET_INFO_KHR,
|
|
.videoSessionParameters = ctx->session_params,
|
|
};
|
|
params_feedback = (VkVideoEncodeSessionParametersFeedbackInfoKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_SESSION_PARAMETERS_FEEDBACK_INFO_KHR,
|
|
};
|
|
|
|
ret = vk->GetEncodedVideoSessionParametersKHR(s->hwctx->act_dev, ¶ms_info,
|
|
¶ms_feedback,
|
|
&data_size, data);
|
|
if (ret == VK_INCOMPLETE ||
|
|
(ret == VK_SUCCESS) && (data_size > 0)) {
|
|
data = av_mallocz(data_size);
|
|
if (!data)
|
|
return AVERROR(ENOMEM);
|
|
} else {
|
|
av_log(avctx, AV_LOG_ERROR, "Unable to get feedback for AV1 sequence header = %"SIZE_SPECIFIER"\n",
|
|
data_size);
|
|
return err;
|
|
}
|
|
|
|
ret = vk->GetEncodedVideoSessionParametersKHR(s->hwctx->act_dev, ¶ms_info,
|
|
¶ms_feedback,
|
|
&data_size, data);
|
|
if (ret != VK_SUCCESS) {
|
|
av_log(avctx, AV_LOG_ERROR, "Error writing feedback units\n");
|
|
return err;
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "Feedback units written, overrides: %i\n",
|
|
params_feedback.hasOverrides);
|
|
|
|
params_feedback.hasOverrides = 1;
|
|
|
|
/* No need to sync any overrides */
|
|
if (!params_feedback.hasOverrides)
|
|
return 0;
|
|
|
|
/* Parse back tne units and override */
|
|
err = parse_feedback_units(avctx, data, data_size);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
/* Create final session parameters */
|
|
err = create_session_params(avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int vulkan_encode_av1_add_obu(AVCodecContext *avctx,
|
|
CodedBitstreamFragment *au,
|
|
uint8_t type, void *obu_unit)
|
|
{
|
|
int err;
|
|
|
|
err = ff_cbs_insert_unit_content(au, -1,
|
|
type, obu_unit, NULL);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to add OBU unit: "
|
|
"type = %d.\n", type);
|
|
return err;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int vulkan_encode_av1_write_obu(AVCodecContext *avctx,
|
|
uint8_t *data, size_t *data_len,
|
|
CodedBitstreamFragment *obu)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
int ret;
|
|
|
|
ret = ff_cbs_write_fragment_data(enc->cbs, obu);
|
|
if (ret < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to write packed header.\n");
|
|
return ret;
|
|
}
|
|
|
|
memcpy(data, obu->data, obu->data_size);
|
|
*data_len = obu->data_size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int write_sequence_header(AVCodecContext *avctx,
|
|
FFHWBaseEncodePicture *base_pic,
|
|
uint8_t *data, size_t *data_len)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
CodedBitstreamFragment *obu = &enc->current_access_unit;
|
|
|
|
err = vulkan_encode_av1_add_obu(avctx, obu,
|
|
AV1_OBU_SEQUENCE_HEADER, &enc->seq_hdr_obu);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vulkan_encode_av1_write_obu(avctx, data, data_len, obu);
|
|
|
|
fail:
|
|
ff_cbs_fragment_reset(obu);
|
|
return err;
|
|
}
|
|
|
|
static int write_extra_headers(AVCodecContext *avctx,
|
|
FFHWBaseEncodePicture *base_pic,
|
|
uint8_t *data, size_t *data_len)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
VulkanEncodeAV1Picture *ap = base_pic->codec_priv;
|
|
CodedBitstreamFragment *obu = &enc->current_access_unit;
|
|
|
|
if (ap->units_needed & AV_FRAME_DATA_MASTERING_DISPLAY_METADATA) {
|
|
err = vulkan_encode_av1_add_obu(avctx, obu,
|
|
AV1_OBU_METADATA,
|
|
&enc->meta_mastering_obu);
|
|
if (err < 0)
|
|
goto fail;
|
|
}
|
|
|
|
if (ap->units_needed & UNIT_CONTENT_LIGHT_LEVEL) {
|
|
err = vulkan_encode_av1_add_obu(avctx, obu,
|
|
AV1_OBU_METADATA,
|
|
&enc->meta_cll_obu);
|
|
if (err < 0)
|
|
goto fail;
|
|
}
|
|
|
|
if (ap->units_needed) {
|
|
err = vulkan_encode_av1_write_obu(avctx, data, data_len, obu);
|
|
if (err < 0)
|
|
goto fail;
|
|
} else {
|
|
err = 0;
|
|
*data_len = 0;
|
|
}
|
|
|
|
fail:
|
|
ff_cbs_fragment_reset(obu);
|
|
return err;
|
|
}
|
|
|
|
static int write_padding(AVCodecContext *avctx, uint32_t padding,
|
|
uint8_t *data, size_t *data_len)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
CodedBitstreamFragment *obu = &enc->current_access_unit;
|
|
|
|
AV1RawOBU padding_obu = { 0 };
|
|
AV1RawPadding *raw_padding = &padding_obu.obu.padding;
|
|
|
|
if (!padding)
|
|
padding = 16;
|
|
|
|
/* 2 byte header + 1 byte trailing bits */
|
|
padding_obu.header.obu_type = AV1_OBU_PADDING;
|
|
*raw_padding = (AV1RawPadding) {
|
|
.payload = enc->padding_payload,
|
|
.payload_size = padding,
|
|
};
|
|
|
|
err = vulkan_encode_av1_add_obu(avctx, obu, AV1_OBU_PADDING, &padding_obu);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = vulkan_encode_av1_write_obu(avctx, data, data_len, obu);
|
|
fail:
|
|
ff_cbs_fragment_reset(obu);
|
|
return err;
|
|
}
|
|
|
|
static const FFVulkanCodec enc_cb = {
|
|
.flags = FF_HW_FLAG_B_PICTURES |
|
|
FF_HW_FLAG_B_PICTURE_REFERENCES |
|
|
VK_ENC_FLAG_NO_DELAY |
|
|
FF_HW_FLAG_SLICE_CONTROL,
|
|
.picture_priv_data_size = sizeof(VulkanEncodeAV1Picture),
|
|
.filler_header_size = 4,
|
|
.init_profile = init_profile,
|
|
.init_pic_rc = init_pic_rc,
|
|
.init_pic_params = init_pic_params,
|
|
.write_sequence_headers = write_sequence_header,
|
|
.write_extra_headers = write_extra_headers,
|
|
.write_filler = write_padding,
|
|
};
|
|
|
|
static av_cold int vulkan_encode_av1_init(AVCodecContext *avctx)
|
|
{
|
|
int err;
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
FFVulkanEncodeContext *ctx = &enc->common;
|
|
FFHWBaseEncodeContext *base_ctx = &ctx->base;
|
|
int flags;
|
|
|
|
if (avctx->profile == AV_PROFILE_UNKNOWN)
|
|
avctx->profile = enc->common.opts.profile;
|
|
|
|
enc->caps = (VkVideoEncodeAV1CapabilitiesKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_CAPABILITIES_KHR,
|
|
};
|
|
|
|
enc->quality_props = (VkVideoEncodeAV1QualityLevelPropertiesKHR) {
|
|
.sType = VK_STRUCTURE_TYPE_VIDEO_ENCODE_AV1_QUALITY_LEVEL_PROPERTIES_KHR,
|
|
};
|
|
|
|
err = ff_vulkan_encode_init(avctx, &enc->common,
|
|
&ff_vk_enc_av1_desc, &enc_cb,
|
|
&enc->caps, &enc->quality_props);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "AV1 encoder capabilities:\n");
|
|
av_log(avctx, AV_LOG_VERBOSE, " Standard capability flags:\n");
|
|
av_log(avctx, AV_LOG_VERBOSE, " per_rate_control_group_min_max_q_index: %i\n",
|
|
!!(enc->caps.flags & VK_VIDEO_ENCODE_AV1_CAPABILITY_PER_RATE_CONTROL_GROUP_MIN_MAX_Q_INDEX_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " generate_obu_extension_header: %i\n",
|
|
!!(enc->caps.flags & VK_VIDEO_ENCODE_AV1_CAPABILITY_GENERATE_OBU_EXTENSION_HEADER_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " primary_reference_cdf_only: %i\n",
|
|
!!(enc->caps.flags & VK_VIDEO_ENCODE_AV1_CAPABILITY_PRIMARY_REFERENCE_CDF_ONLY_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " frame_size_override: %i\n",
|
|
!!(enc->caps.flags & VK_VIDEO_ENCODE_AV1_CAPABILITY_FRAME_SIZE_OVERRIDE_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " motion_vector_scaling: %i\n",
|
|
!!(enc->caps.flags & VK_VIDEO_ENCODE_AV1_CAPABILITY_MOTION_VECTOR_SCALING_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " Capabilities:\n");
|
|
av_log(avctx, AV_LOG_VERBOSE, " 64x64 superblocks: %i\n",
|
|
!!(enc->caps.superblockSizes & VK_VIDEO_ENCODE_AV1_SUPERBLOCK_SIZE_64_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " 128x128 superblocks: %i\n",
|
|
!!(enc->caps.superblockSizes & VK_VIDEO_ENCODE_AV1_SUPERBLOCK_SIZE_128_BIT_KHR));
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxSingleReferenceCount: %i\n",
|
|
enc->caps.maxSingleReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " singleReferenceNameMask: 0x%x\n",
|
|
enc->caps.singleReferenceNameMask);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxUnidirectionalCompoundReferenceCount: %i\n",
|
|
enc->caps.maxUnidirectionalCompoundReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxUnidirectionalCompoundGroup1ReferenceCount: %i\n",
|
|
enc->caps.maxUnidirectionalCompoundGroup1ReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " unidirectionalCompoundReferenceNameMask: 0x%x\n",
|
|
enc->caps.unidirectionalCompoundReferenceNameMask);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxBidirectionalCompoundReferenceCount: %i\n",
|
|
enc->caps.maxBidirectionalCompoundReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxBidirectionalCompoundGroup1ReferenceCount: %i\n",
|
|
enc->caps.maxBidirectionalCompoundGroup1ReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxBidirectionalCompoundGroup2ReferenceCount: %i\n",
|
|
enc->caps.maxBidirectionalCompoundGroup2ReferenceCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " bidirectionalCompoundReferenceNameMask: 0x%x\n",
|
|
enc->caps.bidirectionalCompoundReferenceNameMask);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxTemporalLayerCount: %i\n",
|
|
enc->caps.maxTemporalLayerCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxSpatialLayerCount: %i\n",
|
|
enc->caps.maxSpatialLayerCount);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxOperatingPoints: %i\n",
|
|
enc->caps.maxOperatingPoints);
|
|
av_log(avctx, AV_LOG_VERBOSE, " min/max Qindex: [%i, %i]\n",
|
|
enc->caps.minQIndex, enc->caps.maxQIndex);
|
|
av_log(avctx, AV_LOG_VERBOSE, " prefersGopRemainingFrames: %i\n",
|
|
enc->caps.prefersGopRemainingFrames);
|
|
av_log(avctx, AV_LOG_VERBOSE, " requiresGopRemainingFrames: %i\n",
|
|
enc->caps.requiresGopRemainingFrames);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxLevel: %i\n",
|
|
enc->caps.maxLevel);
|
|
av_log(avctx, AV_LOG_VERBOSE, " codedPictureAlignment: %ix%i\n",
|
|
enc->caps.codedPictureAlignment.width, enc->caps.codedPictureAlignment.height);
|
|
av_log(avctx, AV_LOG_VERBOSE, " maxTiles: %ix%i\n",
|
|
enc->caps.maxTiles.width, enc->caps.maxTiles.height);
|
|
av_log(avctx, AV_LOG_VERBOSE, " Tile size: %ix%i to %ix%i\n",
|
|
enc->caps.minTileSize.width, enc->caps.minTileSize.height,
|
|
enc->caps.maxTileSize.width, enc->caps.maxTileSize.height);
|
|
|
|
err = init_enc_options(avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
flags = ctx->codec->flags;
|
|
err = ff_hw_base_init_gop_structure(base_ctx, avctx,
|
|
ctx->caps.maxDpbSlots,
|
|
enc->caps.maxBidirectionalCompoundReferenceCount,
|
|
flags, 0);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
base_ctx->output_delay = base_ctx->b_per_p;
|
|
base_ctx->decode_delay = base_ctx->max_b_depth;
|
|
|
|
/* Create units and session parameters */
|
|
err = init_base_units(avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
/* Init CBS */
|
|
err = ff_cbs_init(&enc->cbs, AV_CODEC_ID_AV1, avctx);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
|
|
uint8_t data[4096];
|
|
size_t data_len = sizeof(data);
|
|
|
|
err = write_sequence_header(avctx, NULL, data, &data_len);
|
|
if (err < 0) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed to write sequence header "
|
|
"for extradata: %d.\n", err);
|
|
return err;
|
|
} else {
|
|
avctx->extradata_size = data_len;
|
|
avctx->extradata = av_mallocz(avctx->extradata_size +
|
|
AV_INPUT_BUFFER_PADDING_SIZE);
|
|
if (!avctx->extradata) {
|
|
err = AVERROR(ENOMEM);
|
|
return err;
|
|
}
|
|
memcpy(avctx->extradata, data, avctx->extradata_size);
|
|
}
|
|
}
|
|
|
|
enc->padding_payload = av_mallocz(2*ctx->caps.minBitstreamBufferOffsetAlignment);
|
|
if (!enc->padding_payload)
|
|
return AVERROR(ENOMEM);
|
|
|
|
memset(enc->padding_payload, 0xaa, 2*ctx->caps.minBitstreamBufferOffsetAlignment);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int vulkan_encode_av1_close(AVCodecContext *avctx)
|
|
{
|
|
VulkanEncodeAV1Context *enc = avctx->priv_data;
|
|
av_free(enc->padding_payload);
|
|
ff_vulkan_encode_uninit(&enc->common);
|
|
return 0;
|
|
}
|
|
|
|
#define OFFSET(x) offsetof(VulkanEncodeAV1Context, x)
|
|
#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM)
|
|
static const AVOption vulkan_encode_av1_options[] = {
|
|
HW_BASE_ENCODE_COMMON_OPTIONS,
|
|
VULKAN_ENCODE_COMMON_OPTIONS,
|
|
|
|
{ "profile", "Set profile",
|
|
OFFSET(common.opts.profile), AV_OPT_TYPE_INT,
|
|
{ .i64 = AV_PROFILE_UNKNOWN }, AV_PROFILE_UNKNOWN, 0xffff, FLAGS, .unit = "profile" },
|
|
|
|
#define PROFILE(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
|
|
{ .i64 = value }, 0, 0, FLAGS, .unit = "profile"
|
|
{ PROFILE("main", AV_PROFILE_AV1_MAIN) },
|
|
{ PROFILE("high", AV_PROFILE_AV1_HIGH) },
|
|
{ PROFILE("professional", AV_PROFILE_AV1_PROFESSIONAL) },
|
|
#undef PROFILE
|
|
|
|
{ "tier", "Set tier (seq_tier)",
|
|
OFFSET(common.opts.tier), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 1, FLAGS, .unit = "tier" },
|
|
{ "main", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 0 }, 0, 0, FLAGS, .unit = "tier" },
|
|
{ "high", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 1 }, 0, 0, FLAGS, .unit = "tier" },
|
|
|
|
{ "level", "Set level (level_idc)",
|
|
OFFSET(common.opts.level), AV_OPT_TYPE_INT,
|
|
{ .i64 = AV_LEVEL_UNKNOWN }, AV_LEVEL_UNKNOWN, 0xff, FLAGS, .unit = "level" },
|
|
|
|
#define LEVEL(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
|
|
{ .i64 = value }, 0, 0, FLAGS, .unit = "level"
|
|
{ LEVEL("20", 0) },
|
|
{ LEVEL("21", 1) },
|
|
{ LEVEL("22", 2) },
|
|
{ LEVEL("23", 3) },
|
|
{ LEVEL("30", 4) },
|
|
{ LEVEL("31", 5) },
|
|
{ LEVEL("32", 6) },
|
|
{ LEVEL("33", 7) },
|
|
{ LEVEL("40", 8) },
|
|
{ LEVEL("41", 9) },
|
|
{ LEVEL("42", 10) },
|
|
{ LEVEL("43", 11) },
|
|
{ LEVEL("50", 12) },
|
|
{ LEVEL("51", 13) },
|
|
{ LEVEL("52", 14) },
|
|
{ LEVEL("53", 15) },
|
|
{ LEVEL("60", 16) },
|
|
{ LEVEL("61", 17) },
|
|
{ LEVEL("62", 18) },
|
|
{ LEVEL("63", 19) },
|
|
{ LEVEL("70", 20) },
|
|
{ LEVEL("71", 21) },
|
|
{ LEVEL("72", 22) },
|
|
{ LEVEL("73", 23) },
|
|
#undef LEVEL
|
|
|
|
{ "units", "Set units to include", OFFSET(unit_elems), AV_OPT_TYPE_FLAGS, { .i64 = UNIT_MASTERING_DISPLAY | UNIT_CONTENT_LIGHT_LEVEL }, 0, INT_MAX, FLAGS, "units" },
|
|
{ "hdr", "Include HDR metadata for mastering display colour volume and content light level information", 0, AV_OPT_TYPE_CONST, { .i64 = UNIT_MASTERING_DISPLAY | UNIT_CONTENT_LIGHT_LEVEL }, INT_MIN, INT_MAX, FLAGS, "units" },
|
|
|
|
{ NULL },
|
|
};
|
|
|
|
static const FFCodecDefault vulkan_encode_av1_defaults[] = {
|
|
{ "b", "0" },
|
|
{ "bf", "2" },
|
|
{ "g", "300" },
|
|
{ "qmin", "1" },
|
|
{ "qmax", "255" },
|
|
{ "refs", "0" },
|
|
{ NULL },
|
|
};
|
|
|
|
static const AVClass vulkan_encode_av1_class = {
|
|
.class_name = "av1_vulkan",
|
|
.item_name = av_default_item_name,
|
|
.option = vulkan_encode_av1_options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
const FFCodec ff_av1_vulkan_encoder = {
|
|
.p.name = "av1_vulkan",
|
|
CODEC_LONG_NAME("AV1 (Vulkan)"),
|
|
.p.type = AVMEDIA_TYPE_VIDEO,
|
|
.p.id = AV_CODEC_ID_AV1,
|
|
.priv_data_size = sizeof(VulkanEncodeAV1Context),
|
|
.init = &vulkan_encode_av1_init,
|
|
FF_CODEC_RECEIVE_PACKET_CB(&ff_vulkan_encode_receive_packet),
|
|
.close = &vulkan_encode_av1_close,
|
|
.p.priv_class = &vulkan_encode_av1_class,
|
|
.p.capabilities = AV_CODEC_CAP_DELAY |
|
|
AV_CODEC_CAP_HARDWARE |
|
|
AV_CODEC_CAP_DR1 |
|
|
AV_CODEC_CAP_ENCODER_FLUSH |
|
|
AV_CODEC_CAP_ENCODER_REORDERED_OPAQUE,
|
|
.caps_internal = FF_CODEC_CAP_INIT_CLEANUP,
|
|
.defaults = vulkan_encode_av1_defaults,
|
|
CODEC_PIXFMTS(AV_PIX_FMT_VULKAN),
|
|
.hw_configs = ff_vulkan_encode_hw_configs,
|
|
.p.wrapper_name = "vulkan",
|
|
};
|