1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-28 20:53:54 +02:00
FFmpeg/libavcodec/vaapi_encode_mpeg2.c
Linjie Fu 2b32068916 lavc/vaapi_encode: add FF_CODEC_CAP_INIT_CLEANUP caps for encoders
ff_vaapi_encode_close() is not enough to free the resources like cbs
if initialization failure happens after codec->configure (except for
vp8/vp9).

We need to call avctx->codec->close() to deallocate, otherwise memory
leak happens.

Add FF_CODEC_CAP_INIT_CLEANUP for vaapi encoders and deallocate the
resources at free_and_end inside avcodec_open2().

Reviewed-by: Timo Rothenpieler <timo@rothenpieler.org>
Signed-off-by: Linjie Fu <linjie.fu@intel.com>
2020-05-04 12:33:30 -03:00

714 lines
24 KiB
C

/*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <va/va.h>
#include <va/va_enc_mpeg2.h>
#include "libavutil/avassert.h"
#include "avcodec.h"
#include "cbs.h"
#include "cbs_mpeg2.h"
#include "mpeg12.h"
#include "vaapi_encode.h"
typedef struct VAAPIEncodeMPEG2Context {
VAAPIEncodeContext common;
// User options.
int profile;
int level;
// Derived settings.
int quant_i;
int quant_p;
int quant_b;
unsigned int bit_rate;
unsigned int vbv_buffer_size;
AVRational frame_rate;
unsigned int f_code_horizontal;
unsigned int f_code_vertical;
// Stream state.
int64_t last_i_frame;
// Writer structures.
MPEG2RawSequenceHeader sequence_header;
MPEG2RawExtensionData sequence_extension;
MPEG2RawExtensionData sequence_display_extension;
MPEG2RawGroupOfPicturesHeader gop_header;
MPEG2RawPictureHeader picture_header;
MPEG2RawExtensionData picture_coding_extension;
CodedBitstreamContext *cbc;
CodedBitstreamFragment current_fragment;
} VAAPIEncodeMPEG2Context;
static int vaapi_encode_mpeg2_write_fragment(AVCodecContext *avctx,
char *data, size_t *data_len,
CodedBitstreamFragment *frag)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
int err;
err = ff_cbs_write_fragment_data(priv->cbc, frag);
if (err < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to write packed header.\n");
return err;
}
if (*data_len < 8 * frag->data_size - frag->data_bit_padding) {
av_log(avctx, AV_LOG_ERROR, "Access unit too large: "
"%zu < %zu.\n", *data_len,
8 * frag->data_size - frag->data_bit_padding);
return AVERROR(ENOSPC);
}
memcpy(data, frag->data, frag->data_size);
*data_len = 8 * frag->data_size - frag->data_bit_padding;
return 0;
}
static int vaapi_encode_mpeg2_add_header(AVCodecContext *avctx,
CodedBitstreamFragment *frag,
int type, void *header)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
int err;
err = ff_cbs_insert_unit_content(priv->cbc, frag, -1, type, header, NULL);
if (err < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to add header: "
"type = %d.\n", type);
return err;
}
return 0;
}
static int vaapi_encode_mpeg2_write_sequence_header(AVCodecContext *avctx,
char *data, size_t *data_len)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
CodedBitstreamFragment *frag = &priv->current_fragment;
int err;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_SEQUENCE_HEADER,
&priv->sequence_header);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_EXTENSION,
&priv->sequence_extension);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_EXTENSION,
&priv->sequence_display_extension);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_GROUP,
&priv->gop_header);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_write_fragment(avctx, data, data_len, frag);
fail:
ff_cbs_fragment_reset(priv->cbc, frag);
return 0;
}
static int vaapi_encode_mpeg2_write_picture_header(AVCodecContext *avctx,
VAAPIEncodePicture *pic,
char *data, size_t *data_len)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
CodedBitstreamFragment *frag = &priv->current_fragment;
int err;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_PICTURE,
&priv->picture_header);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_add_header(avctx, frag, MPEG2_START_EXTENSION,
&priv->picture_coding_extension);
if (err < 0)
goto fail;
err = vaapi_encode_mpeg2_write_fragment(avctx, data, data_len, frag);
fail:
ff_cbs_fragment_reset(priv->cbc, frag);
return 0;
}
static int vaapi_encode_mpeg2_init_sequence_params(AVCodecContext *avctx)
{
VAAPIEncodeContext *ctx = avctx->priv_data;
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
MPEG2RawSequenceHeader *sh = &priv->sequence_header;
MPEG2RawSequenceExtension *se = &priv->sequence_extension.data.sequence;
MPEG2RawSequenceDisplayExtension *sde = &priv->sequence_display_extension.data.sequence_display;
MPEG2RawGroupOfPicturesHeader *goph = &priv->gop_header;
MPEG2RawPictureHeader *ph = &priv->picture_header;
MPEG2RawPictureCodingExtension *pce = &priv->picture_coding_extension.data.picture_coding;
VAEncSequenceParameterBufferMPEG2 *vseq = ctx->codec_sequence_params;
VAEncPictureParameterBufferMPEG2 *vpic = ctx->codec_picture_params;
int code, ext_n, ext_d;
memset(sh, 0, sizeof(*sh));
memset(se, 0, sizeof(*se));
memset(sde, 0, sizeof(*sde));
memset(goph, 0, sizeof(*goph));
memset(ph, 0, sizeof(*ph));
memset(pce, 0, sizeof(*pce));
if (ctx->va_bit_rate > 0) {
priv->bit_rate = (ctx->va_bit_rate + 399) / 400;
} else {
// Unknown (not a bitrate-targetting mode), so just use the
// highest value.
priv->bit_rate = 0x3fffffff;
}
if (avctx->rc_buffer_size > 0) {
priv->vbv_buffer_size = (avctx->rc_buffer_size + (1 << 14) - 1) >> 14;
} else {
// Unknown, so guess a value from the bitrate.
priv->vbv_buffer_size = priv->bit_rate >> 14;
}
switch (avctx->level) {
case 4: // High.
case 6: // High 1440.
priv->f_code_horizontal = 9;
priv->f_code_vertical = 5;
break;
case 8: // Main.
priv->f_code_horizontal = 8;
priv->f_code_vertical = 5;
break;
case 10: // Low.
default:
priv->f_code_horizontal = 7;
priv->f_code_vertical = 4;
break;
}
// Sequence header
sh->sequence_header_code = MPEG2_START_SEQUENCE_HEADER;
sh->horizontal_size_value = avctx->width & 0xfff;
sh->vertical_size_value = avctx->height & 0xfff;
if (avctx->sample_aspect_ratio.num != 0 &&
avctx->sample_aspect_ratio.den != 0) {
AVRational dar = av_div_q(avctx->sample_aspect_ratio,
(AVRational) { avctx->width, avctx->height });
if (av_cmp_q(avctx->sample_aspect_ratio, (AVRational) { 1, 1 }) == 0) {
sh->aspect_ratio_information = 1;
} else if (av_cmp_q(dar, (AVRational) { 3, 4 }) == 0) {
sh->aspect_ratio_information = 2;
} else if (av_cmp_q(dar, (AVRational) { 9, 16 }) == 0) {
sh->aspect_ratio_information = 3;
} else if (av_cmp_q(dar, (AVRational) { 100, 221 }) == 0) {
sh->aspect_ratio_information = 4;
} else {
av_log(avctx, AV_LOG_WARNING, "Sample aspect ratio %d:%d is not "
"representable, signalling square pixels instead.\n",
avctx->sample_aspect_ratio.num,
avctx->sample_aspect_ratio.den);
sh->aspect_ratio_information = 1;
}
} else {
// Unknown - assume square pixels.
sh->aspect_ratio_information = 1;
}
if (avctx->framerate.num > 0 && avctx->framerate.den > 0)
priv->frame_rate = avctx->framerate;
else
priv->frame_rate = av_inv_q(avctx->time_base);
ff_mpeg12_find_best_frame_rate(priv->frame_rate,
&code, &ext_n, &ext_d, 0);
sh->frame_rate_code = code;
sh->bit_rate_value = priv->bit_rate & 0x3ffff;
sh->vbv_buffer_size_value = priv->vbv_buffer_size & 0x3ff;
sh->constrained_parameters_flag = 0;
sh->load_intra_quantiser_matrix = 0;
sh->load_non_intra_quantiser_matrix = 0;
// Sequence extension
priv->sequence_extension.extension_start_code = MPEG2_START_EXTENSION;
priv->sequence_extension.extension_start_code_identifier =
MPEG2_EXTENSION_SEQUENCE;
se->profile_and_level_indication = avctx->profile << 4 | avctx->level;
se->progressive_sequence = 1;
se->chroma_format = 1;
se->horizontal_size_extension = avctx->width >> 12;
se->vertical_size_extension = avctx->height >> 12;
se->bit_rate_extension = priv->bit_rate >> 18;
se->vbv_buffer_size_extension = priv->vbv_buffer_size >> 10;
se->low_delay = ctx->b_per_p == 0;
se->frame_rate_extension_n = ext_n;
se->frame_rate_extension_d = ext_d;
// Sequence display extension
priv->sequence_display_extension.extension_start_code =
MPEG2_START_EXTENSION;
priv->sequence_display_extension.extension_start_code_identifier =
MPEG2_EXTENSION_SEQUENCE_DISPLAY;
sde->video_format = 5;
if (avctx->color_primaries != AVCOL_PRI_UNSPECIFIED ||
avctx->color_trc != AVCOL_TRC_UNSPECIFIED ||
avctx->colorspace != AVCOL_SPC_UNSPECIFIED) {
sde->colour_description = 1;
sde->colour_primaries = avctx->color_primaries;
sde->transfer_characteristics = avctx->color_trc;
sde->matrix_coefficients = avctx->colorspace;
} else {
sde->colour_description = 0;
}
sde->display_horizontal_size = avctx->width;
sde->display_vertical_size = avctx->height;
// GOP header
goph->group_start_code = MPEG2_START_GROUP;
// Marker bit in the middle of time_code.
goph->time_code = 1 << 12;
goph->closed_gop = 1;
goph->broken_link = 0;
// Defaults for picture header
ph->picture_start_code = MPEG2_START_PICTURE;
ph->vbv_delay = 0xffff; // Not currently calculated.
ph->full_pel_forward_vector = 0;
ph->forward_f_code = 7;
ph->full_pel_backward_vector = 0;
ph->forward_f_code = 7;
// Defaults for picture coding extension
priv->picture_coding_extension.extension_start_code =
MPEG2_START_EXTENSION;
priv->picture_coding_extension.extension_start_code_identifier =
MPEG2_EXTENSION_PICTURE_CODING;
pce->intra_dc_precision = 0;
pce->picture_structure = 3;
pce->top_field_first = 0;
pce->frame_pred_frame_dct = 1;
pce->concealment_motion_vectors = 0;
pce->q_scale_type = 0;
pce->intra_vlc_format = 0;
pce->alternate_scan = 0;
pce->repeat_first_field = 0;
pce->progressive_frame = 1;
pce->composite_display_flag = 0;
*vseq = (VAEncSequenceParameterBufferMPEG2) {
.intra_period = ctx->gop_size,
.ip_period = ctx->b_per_p + 1,
.picture_width = avctx->width,
.picture_height = avctx->height,
.bits_per_second = ctx->va_bit_rate,
.frame_rate = av_q2d(priv->frame_rate),
.aspect_ratio_information = sh->aspect_ratio_information,
.vbv_buffer_size = priv->vbv_buffer_size,
.sequence_extension.bits = {
.profile_and_level_indication = se->profile_and_level_indication,
.progressive_sequence = se->progressive_sequence,
.chroma_format = se->chroma_format,
.low_delay = se->low_delay,
.frame_rate_extension_n = se->frame_rate_extension_n,
.frame_rate_extension_d = se->frame_rate_extension_d,
},
.new_gop_header = 1,
.gop_header.bits = {
.time_code = goph->time_code,
.closed_gop = goph->closed_gop,
.broken_link = goph->broken_link,
},
};
*vpic = (VAEncPictureParameterBufferMPEG2) {
.forward_reference_picture = VA_INVALID_ID,
.backward_reference_picture = VA_INVALID_ID,
.reconstructed_picture = VA_INVALID_ID,
.coded_buf = VA_INVALID_ID,
.vbv_delay = 0xffff,
.f_code = { { 15, 15 }, { 15, 15 } },
.picture_coding_extension.bits = {
.intra_dc_precision = pce->intra_dc_precision,
.picture_structure = pce->picture_structure,
.top_field_first = pce->top_field_first,
.frame_pred_frame_dct = pce->frame_pred_frame_dct,
.concealment_motion_vectors = pce->concealment_motion_vectors,
.q_scale_type = pce->q_scale_type,
.intra_vlc_format = pce->intra_vlc_format,
.alternate_scan = pce->alternate_scan,
.repeat_first_field = pce->repeat_first_field,
.progressive_frame = pce->progressive_frame,
.composite_display_flag = pce->composite_display_flag,
},
.composite_display.bits = {
.v_axis = pce->v_axis,
.field_sequence = pce->field_sequence,
.sub_carrier = pce->sub_carrier,
.burst_amplitude = pce->burst_amplitude,
.sub_carrier_phase = pce->sub_carrier_phase,
},
};
return 0;
}
static int vaapi_encode_mpeg2_init_picture_params(AVCodecContext *avctx,
VAAPIEncodePicture *pic)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
MPEG2RawPictureHeader *ph = &priv->picture_header;
MPEG2RawPictureCodingExtension *pce = &priv->picture_coding_extension.data.picture_coding;
VAEncPictureParameterBufferMPEG2 *vpic = pic->codec_picture_params;
if (pic->type == PICTURE_TYPE_IDR || pic->type == PICTURE_TYPE_I) {
ph->temporal_reference = 0;
ph->picture_coding_type = 1;
priv->last_i_frame = pic->display_order;
} else {
ph->temporal_reference = pic->display_order - priv->last_i_frame;
ph->picture_coding_type = pic->type == PICTURE_TYPE_B ? 3 : 2;
}
if (pic->type == PICTURE_TYPE_P || pic->type == PICTURE_TYPE_B) {
pce->f_code[0][0] = priv->f_code_horizontal;
pce->f_code[0][1] = priv->f_code_vertical;
} else {
pce->f_code[0][0] = 15;
pce->f_code[0][1] = 15;
}
if (pic->type == PICTURE_TYPE_B) {
pce->f_code[1][0] = priv->f_code_horizontal;
pce->f_code[1][1] = priv->f_code_vertical;
} else {
pce->f_code[1][0] = 15;
pce->f_code[1][1] = 15;
}
vpic->reconstructed_picture = pic->recon_surface;
vpic->coded_buf = pic->output_buffer;
switch (pic->type) {
case PICTURE_TYPE_IDR:
case PICTURE_TYPE_I:
vpic->picture_type = VAEncPictureTypeIntra;
break;
case PICTURE_TYPE_P:
vpic->picture_type = VAEncPictureTypePredictive;
vpic->forward_reference_picture = pic->refs[0]->recon_surface;
break;
case PICTURE_TYPE_B:
vpic->picture_type = VAEncPictureTypeBidirectional;
vpic->forward_reference_picture = pic->refs[0]->recon_surface;
vpic->backward_reference_picture = pic->refs[1]->recon_surface;
break;
default:
av_assert0(0 && "invalid picture type");
}
vpic->temporal_reference = ph->temporal_reference;
vpic->f_code[0][0] = pce->f_code[0][0];
vpic->f_code[0][1] = pce->f_code[0][1];
vpic->f_code[1][0] = pce->f_code[1][0];
vpic->f_code[1][1] = pce->f_code[1][1];
return 0;
}
static int vaapi_encode_mpeg2_init_slice_params(AVCodecContext *avctx,
VAAPIEncodePicture *pic,
VAAPIEncodeSlice *slice)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
VAEncSliceParameterBufferMPEG2 *vslice = slice->codec_slice_params;
int qp;
vslice->macroblock_address = slice->block_start;
vslice->num_macroblocks = slice->block_size;
switch (pic->type) {
case PICTURE_TYPE_IDR:
case PICTURE_TYPE_I:
qp = priv->quant_i;
break;
case PICTURE_TYPE_P:
qp = priv->quant_p;
break;
case PICTURE_TYPE_B:
qp = priv->quant_b;
break;
default:
av_assert0(0 && "invalid picture type");
}
vslice->quantiser_scale_code = qp;
vslice->is_intra_slice = (pic->type == PICTURE_TYPE_IDR ||
pic->type == PICTURE_TYPE_I);
return 0;
}
static av_cold int vaapi_encode_mpeg2_configure(AVCodecContext *avctx)
{
VAAPIEncodeContext *ctx = avctx->priv_data;
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
int err;
err = ff_cbs_init(&priv->cbc, AV_CODEC_ID_MPEG2VIDEO, avctx);
if (err < 0)
return err;
if (ctx->va_rc_mode == VA_RC_CQP) {
priv->quant_p = av_clip(ctx->rc_quality, 1, 31);
if (avctx->i_quant_factor > 0.0)
priv->quant_i =
av_clip((avctx->i_quant_factor * priv->quant_p +
avctx->i_quant_offset) + 0.5, 1, 31);
else
priv->quant_i = priv->quant_p;
if (avctx->b_quant_factor > 0.0)
priv->quant_b =
av_clip((avctx->b_quant_factor * priv->quant_p +
avctx->b_quant_offset) + 0.5, 1, 31);
else
priv->quant_b = priv->quant_p;
av_log(avctx, AV_LOG_DEBUG, "Using fixed quantiser "
"%d / %d / %d for I- / P- / B-frames.\n",
priv->quant_i, priv->quant_p, priv->quant_b);
} else {
priv->quant_i = 16;
priv->quant_p = 16;
priv->quant_b = 16;
}
ctx->slice_block_rows = FFALIGN(avctx->height, 16) / 16;
ctx->slice_block_cols = FFALIGN(avctx->width, 16) / 16;
ctx->nb_slices = ctx->slice_block_rows;
ctx->slice_size = 1;
ctx->roi_quant_range = 31;
return 0;
}
static const VAAPIEncodeProfile vaapi_encode_mpeg2_profiles[] = {
{ FF_PROFILE_MPEG2_MAIN, 8, 3, 1, 1, VAProfileMPEG2Main },
{ FF_PROFILE_MPEG2_SIMPLE, 8, 3, 1, 1, VAProfileMPEG2Simple },
{ FF_PROFILE_UNKNOWN }
};
static const VAAPIEncodeType vaapi_encode_type_mpeg2 = {
.profiles = vaapi_encode_mpeg2_profiles,
.flags = FLAG_B_PICTURES,
.configure = &vaapi_encode_mpeg2_configure,
.default_quality = 10,
.sequence_params_size = sizeof(VAEncSequenceParameterBufferMPEG2),
.init_sequence_params = &vaapi_encode_mpeg2_init_sequence_params,
.picture_params_size = sizeof(VAEncPictureParameterBufferMPEG2),
.init_picture_params = &vaapi_encode_mpeg2_init_picture_params,
.slice_params_size = sizeof(VAEncSliceParameterBufferMPEG2),
.init_slice_params = &vaapi_encode_mpeg2_init_slice_params,
.sequence_header_type = VAEncPackedHeaderSequence,
.write_sequence_header = &vaapi_encode_mpeg2_write_sequence_header,
.picture_header_type = VAEncPackedHeaderPicture,
.write_picture_header = &vaapi_encode_mpeg2_write_picture_header,
};
static av_cold int vaapi_encode_mpeg2_init(AVCodecContext *avctx)
{
VAAPIEncodeContext *ctx = avctx->priv_data;
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
ctx->codec = &vaapi_encode_type_mpeg2;
if (avctx->profile == FF_PROFILE_UNKNOWN)
avctx->profile = priv->profile;
if (avctx->level == FF_LEVEL_UNKNOWN)
avctx->level = priv->level;
// Reject unknown levels (these are required to set f_code for
// motion vector encoding).
switch (avctx->level) {
case 4: // High
case 6: // High 1440
case 8: // Main
case 10: // Low
break;
default:
av_log(avctx, AV_LOG_ERROR, "Unknown MPEG-2 level %d.\n",
avctx->level);
return AVERROR(EINVAL);
}
if (avctx->height % 4096 == 0 || avctx->width % 4096 == 0) {
av_log(avctx, AV_LOG_ERROR, "MPEG-2 does not support picture "
"height or width divisible by 4096.\n");
return AVERROR(EINVAL);
}
ctx->desired_packed_headers = VA_ENC_PACKED_HEADER_SEQUENCE |
VA_ENC_PACKED_HEADER_PICTURE;
ctx->surface_width = FFALIGN(avctx->width, 16);
ctx->surface_height = FFALIGN(avctx->height, 16);
return ff_vaapi_encode_init(avctx);
}
static av_cold int vaapi_encode_mpeg2_close(AVCodecContext *avctx)
{
VAAPIEncodeMPEG2Context *priv = avctx->priv_data;
ff_cbs_fragment_free(priv->cbc, &priv->current_fragment);
ff_cbs_close(&priv->cbc);
return ff_vaapi_encode_close(avctx);
}
#define OFFSET(x) offsetof(VAAPIEncodeMPEG2Context, x)
#define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM)
static const AVOption vaapi_encode_mpeg2_options[] = {
VAAPI_ENCODE_COMMON_OPTIONS,
VAAPI_ENCODE_RC_OPTIONS,
{ "profile", "Set profile (in profile_and_level_indication)",
OFFSET(profile), AV_OPT_TYPE_INT,
{ .i64 = FF_PROFILE_UNKNOWN }, FF_PROFILE_UNKNOWN, 7, FLAGS, "profile" },
#define PROFILE(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
{ .i64 = value }, 0, 0, FLAGS, "profile"
{ PROFILE("simple", FF_PROFILE_MPEG2_SIMPLE) },
{ PROFILE("main", FF_PROFILE_MPEG2_MAIN) },
#undef PROFILE
{ "level", "Set level (in profile_and_level_indication)",
OFFSET(level), AV_OPT_TYPE_INT,
{ .i64 = 4 }, 0, 15, FLAGS, "level" },
#define LEVEL(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \
{ .i64 = value }, 0, 0, FLAGS, "level"
{ LEVEL("low", 10) },
{ LEVEL("main", 8) },
{ LEVEL("high_1440", 6) },
{ LEVEL("high", 4) },
#undef LEVEL
{ NULL },
};
static const AVCodecDefault vaapi_encode_mpeg2_defaults[] = {
{ "b", "0" },
{ "bf", "1" },
{ "g", "120" },
{ "i_qfactor", "1" },
{ "i_qoffset", "0" },
{ "b_qfactor", "6/5" },
{ "b_qoffset", "0" },
{ "qmin", "-1" },
{ "qmax", "-1" },
{ NULL },
};
static const AVClass vaapi_encode_mpeg2_class = {
.class_name = "mpeg2_vaapi",
.item_name = av_default_item_name,
.option = vaapi_encode_mpeg2_options,
.version = LIBAVUTIL_VERSION_INT,
};
AVCodec ff_mpeg2_vaapi_encoder = {
.name = "mpeg2_vaapi",
.long_name = NULL_IF_CONFIG_SMALL("MPEG-2 (VAAPI)"),
.type = AVMEDIA_TYPE_VIDEO,
.id = AV_CODEC_ID_MPEG2VIDEO,
.priv_data_size = sizeof(VAAPIEncodeMPEG2Context),
.init = &vaapi_encode_mpeg2_init,
.send_frame = &ff_vaapi_encode_send_frame,
.receive_packet = &ff_vaapi_encode_receive_packet,
.close = &vaapi_encode_mpeg2_close,
.priv_class = &vaapi_encode_mpeg2_class,
.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE,
.caps_internal = FF_CODEC_CAP_INIT_CLEANUP,
.defaults = vaapi_encode_mpeg2_defaults,
.pix_fmts = (const enum AVPixelFormat[]) {
AV_PIX_FMT_VAAPI,
AV_PIX_FMT_NONE,
},
.hw_configs = ff_vaapi_encode_hw_configs,
.wrapper_name = "vaapi",
};