mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-21 10:55:51 +02:00
avcodec/hw_base_encode: avoid getting FFHWBaseEncodeContext from avctx
This patch is to make FFHWBaseEncodeContext a standalone component and avoid getting FFHWBaseEncodeContext from avctx->priv_data. This patch also removes some unnecessary AVCodecContext arguments. For receive_packet call, a small wrapper is introduced. Signed-off-by: Tong Wu <tong1.wu@intel.com>
This commit is contained in:
parent
0ba10f2d75
commit
e783e45e29
@ -676,6 +676,7 @@ end:
|
||||
static int d3d12va_encode_output(AVCodecContext *avctx,
|
||||
const FFHWBaseEncodePicture *base_pic, AVPacket *pkt)
|
||||
{
|
||||
FFHWBaseEncodeContext *base_ctx = avctx->priv_data;
|
||||
D3D12VAEncodePicture *pic = (D3D12VAEncodePicture *)base_pic;
|
||||
AVPacket *pkt_ptr = pkt;
|
||||
int err;
|
||||
@ -691,7 +692,8 @@ static int d3d12va_encode_output(AVCodecContext *avctx,
|
||||
av_log(avctx, AV_LOG_DEBUG, "Output read for pic %"PRId64"/%"PRId64".\n",
|
||||
base_pic->display_order, base_pic->encode_order);
|
||||
|
||||
ff_hw_base_encode_set_output_property(avctx, (FFHWBaseEncodePicture *)base_pic, pkt_ptr, 0);
|
||||
ff_hw_base_encode_set_output_property(base_ctx, avctx, (FFHWBaseEncodePicture *)base_pic,
|
||||
pkt_ptr, 0);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@ -1119,7 +1121,7 @@ static int d3d12va_encode_init_gop_structure(AVCodecContext *avctx)
|
||||
"replacing them with B-frames.\n");
|
||||
}
|
||||
|
||||
err = ff_hw_base_init_gop_structure(avctx, ref_l0, ref_l1, ctx->codec->flags, 0);
|
||||
err = ff_hw_base_init_gop_structure(base_ctx, avctx, ref_l0, ref_l1, ctx->codec->flags, 0);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@ -1351,7 +1353,7 @@ static int d3d12va_encode_create_recon_frames(AVCodecContext *avctx)
|
||||
enum AVPixelFormat recon_format;
|
||||
int err;
|
||||
|
||||
err = ff_hw_base_get_recon_format(avctx, NULL, &recon_format);
|
||||
err = ff_hw_base_get_recon_format(base_ctx, NULL, &recon_format);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@ -1390,6 +1392,11 @@ static const FFHWEncodePictureOperation d3d12va_type = {
|
||||
.free = &d3d12va_encode_free,
|
||||
};
|
||||
|
||||
int ff_d3d12va_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt)
|
||||
{
|
||||
return ff_hw_base_encode_receive_packet(avctx->priv_data, avctx, pkt);
|
||||
}
|
||||
|
||||
int ff_d3d12va_encode_init(AVCodecContext *avctx)
|
||||
{
|
||||
FFHWBaseEncodeContext *base_ctx = avctx->priv_data;
|
||||
@ -1398,7 +1405,7 @@ int ff_d3d12va_encode_init(AVCodecContext *avctx)
|
||||
int err;
|
||||
HRESULT hr;
|
||||
|
||||
err = ff_hw_base_encode_init(avctx);
|
||||
err = ff_hw_base_encode_init(avctx, base_ctx);
|
||||
if (err < 0)
|
||||
goto fail;
|
||||
|
||||
@ -1552,7 +1559,7 @@ int ff_d3d12va_encode_close(AVCodecContext *avctx)
|
||||
D3D12_OBJECT_RELEASE(ctx->video_device3);
|
||||
D3D12_OBJECT_RELEASE(ctx->device3);
|
||||
|
||||
ff_hw_base_encode_close(avctx);
|
||||
ff_hw_base_encode_close(base_ctx);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -313,6 +313,8 @@ typedef struct D3D12VAEncodeType {
|
||||
char *data, size_t *data_len);
|
||||
} D3D12VAEncodeType;
|
||||
|
||||
int ff_d3d12va_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt);
|
||||
|
||||
int ff_d3d12va_encode_init(AVCodecContext *avctx);
|
||||
int ff_d3d12va_encode_close(AVCodecContext *avctx);
|
||||
|
||||
|
@ -990,7 +990,7 @@ const FFCodec ff_hevc_d3d12va_encoder = {
|
||||
.p.id = AV_CODEC_ID_HEVC,
|
||||
.priv_data_size = sizeof(D3D12VAEncodeHEVCContext),
|
||||
.init = &d3d12va_encode_hevc_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_d3d12va_encode_receive_packet),
|
||||
.close = &d3d12va_encode_hevc_close,
|
||||
.p.priv_class = &d3d12va_encode_hevc_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -94,14 +94,13 @@ static void hw_base_encode_remove_refs(FFHWBaseEncodePicture *pic, int level)
|
||||
pic->ref_removed[level] = 1;
|
||||
}
|
||||
|
||||
static void hw_base_encode_set_b_pictures(AVCodecContext *avctx,
|
||||
static void hw_base_encode_set_b_pictures(FFHWBaseEncodeContext *ctx,
|
||||
FFHWBaseEncodePicture *start,
|
||||
FFHWBaseEncodePicture *end,
|
||||
FFHWBaseEncodePicture *prev,
|
||||
int current_depth,
|
||||
FFHWBaseEncodePicture **last)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
FFHWBaseEncodePicture *pic, *next, *ref;
|
||||
int i, len;
|
||||
|
||||
@ -148,20 +147,19 @@ static void hw_base_encode_set_b_pictures(AVCodecContext *avctx,
|
||||
hw_base_encode_add_ref(pic, ref, 0, 1, 0);
|
||||
|
||||
if (i > 1)
|
||||
hw_base_encode_set_b_pictures(avctx, start, pic, pic,
|
||||
hw_base_encode_set_b_pictures(ctx, start, pic, pic,
|
||||
current_depth + 1, &next);
|
||||
else
|
||||
next = pic;
|
||||
|
||||
hw_base_encode_set_b_pictures(avctx, pic, end, next,
|
||||
hw_base_encode_set_b_pictures(ctx, pic, end, next,
|
||||
current_depth + 1, last);
|
||||
}
|
||||
}
|
||||
|
||||
static void hw_base_encode_add_next_prev(AVCodecContext *avctx,
|
||||
static void hw_base_encode_add_next_prev(FFHWBaseEncodeContext *ctx,
|
||||
FFHWBaseEncodePicture *pic)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
int i;
|
||||
|
||||
if (!pic)
|
||||
@ -192,9 +190,9 @@ static void hw_base_encode_add_next_prev(AVCodecContext *avctx,
|
||||
}
|
||||
|
||||
static int hw_base_encode_pick_next(AVCodecContext *avctx,
|
||||
FFHWBaseEncodeContext *ctx,
|
||||
FFHWBaseEncodePicture **pic_out)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
FFHWBaseEncodePicture *pic = NULL, *prev = NULL, *next, *start;
|
||||
int i, b_counter, closed_gop_end;
|
||||
|
||||
@ -333,19 +331,18 @@ static int hw_base_encode_pick_next(AVCodecContext *avctx,
|
||||
}
|
||||
|
||||
if (b_counter > 0) {
|
||||
hw_base_encode_set_b_pictures(avctx, start, pic, pic, 1,
|
||||
hw_base_encode_set_b_pictures(ctx, start, pic, pic, 1,
|
||||
&prev);
|
||||
} else {
|
||||
prev = pic;
|
||||
}
|
||||
hw_base_encode_add_next_prev(avctx, prev);
|
||||
hw_base_encode_add_next_prev(ctx, prev);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int hw_base_encode_clear_old(AVCodecContext *avctx)
|
||||
static int hw_base_encode_clear_old(AVCodecContext *avctx, FFHWBaseEncodeContext *ctx)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
FFHWBaseEncodePicture *pic, *prev, *next;
|
||||
|
||||
av_assert0(ctx->pic_start);
|
||||
@ -381,14 +378,12 @@ static int hw_base_encode_clear_old(AVCodecContext *avctx)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int hw_base_encode_check_frame(AVCodecContext *avctx,
|
||||
static int hw_base_encode_check_frame(FFHWBaseEncodeContext *ctx,
|
||||
const AVFrame *frame)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
|
||||
if ((frame->crop_top || frame->crop_bottom ||
|
||||
frame->crop_left || frame->crop_right) && !ctx->crop_warned) {
|
||||
av_log(avctx, AV_LOG_WARNING, "Cropping information on input "
|
||||
av_log(ctx->log_ctx, AV_LOG_WARNING, "Cropping information on input "
|
||||
"frames ignored due to lack of API support.\n");
|
||||
ctx->crop_warned = 1;
|
||||
}
|
||||
@ -398,7 +393,7 @@ static int hw_base_encode_check_frame(AVCodecContext *avctx,
|
||||
av_frame_get_side_data(frame, AV_FRAME_DATA_REGIONS_OF_INTEREST);
|
||||
|
||||
if (sd && !ctx->roi_warned) {
|
||||
av_log(avctx, AV_LOG_WARNING, "ROI side data on input "
|
||||
av_log(ctx->log_ctx, AV_LOG_WARNING, "ROI side data on input "
|
||||
"frames ignored due to lack of driver support.\n");
|
||||
ctx->roi_warned = 1;
|
||||
}
|
||||
@ -407,9 +402,9 @@ static int hw_base_encode_check_frame(AVCodecContext *avctx,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int hw_base_encode_send_frame(AVCodecContext *avctx, AVFrame *frame)
|
||||
static int hw_base_encode_send_frame(AVCodecContext *avctx, FFHWBaseEncodeContext *ctx,
|
||||
AVFrame *frame)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
FFHWBaseEncodePicture *pic;
|
||||
int err;
|
||||
|
||||
@ -417,7 +412,7 @@ static int hw_base_encode_send_frame(AVCodecContext *avctx, AVFrame *frame)
|
||||
av_log(avctx, AV_LOG_DEBUG, "Input frame: %ux%u (%"PRId64").\n",
|
||||
frame->width, frame->height, frame->pts);
|
||||
|
||||
err = hw_base_encode_check_frame(avctx, frame);
|
||||
err = hw_base_encode_check_frame(ctx, frame);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@ -488,12 +483,11 @@ fail:
|
||||
return err;
|
||||
}
|
||||
|
||||
int ff_hw_base_encode_set_output_property(AVCodecContext *avctx,
|
||||
int ff_hw_base_encode_set_output_property(FFHWBaseEncodeContext *ctx,
|
||||
AVCodecContext *avctx,
|
||||
FFHWBaseEncodePicture *pic,
|
||||
AVPacket *pkt, int flag_no_delay)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
|
||||
if (pic->type == FF_HW_PICTURE_TYPE_IDR)
|
||||
pkt->flags |= AV_PKT_FLAG_KEY;
|
||||
|
||||
@ -528,9 +522,9 @@ int ff_hw_base_encode_set_output_property(AVCodecContext *avctx,
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ff_hw_base_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt)
|
||||
int ff_hw_base_encode_receive_packet(FFHWBaseEncodeContext *ctx,
|
||||
AVCodecContext *avctx, AVPacket *pkt)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
FFHWBaseEncodePicture *pic = NULL;
|
||||
AVFrame *frame = ctx->frame;
|
||||
int err;
|
||||
@ -558,7 +552,7 @@ start:
|
||||
if (err == AVERROR_EOF)
|
||||
frame = NULL;
|
||||
|
||||
err = hw_base_encode_send_frame(avctx, frame);
|
||||
err = hw_base_encode_send_frame(avctx, ctx, frame);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@ -571,7 +565,7 @@ start:
|
||||
|
||||
if (ctx->async_encode) {
|
||||
if (av_fifo_can_write(ctx->encode_fifo)) {
|
||||
err = hw_base_encode_pick_next(avctx, &pic);
|
||||
err = hw_base_encode_pick_next(avctx, ctx, &pic);
|
||||
if (!err) {
|
||||
av_assert0(pic);
|
||||
pic->encode_order = ctx->encode_order +
|
||||
@ -596,7 +590,7 @@ start:
|
||||
av_fifo_read(ctx->encode_fifo, &pic, 1);
|
||||
ctx->encode_order = pic->encode_order + 1;
|
||||
} else {
|
||||
err = hw_base_encode_pick_next(avctx, &pic);
|
||||
err = hw_base_encode_pick_next(avctx, ctx, &pic);
|
||||
if (err < 0)
|
||||
return err;
|
||||
av_assert0(pic);
|
||||
@ -619,7 +613,7 @@ start:
|
||||
}
|
||||
|
||||
ctx->output_order = pic->encode_order;
|
||||
hw_base_encode_clear_old(avctx);
|
||||
hw_base_encode_clear_old(avctx, ctx);
|
||||
|
||||
/** loop to get an available pkt in encoder flushing. */
|
||||
if (ctx->end_of_stream && !pkt->size)
|
||||
@ -633,11 +627,10 @@ end:
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ff_hw_base_init_gop_structure(AVCodecContext *avctx, uint32_t ref_l0, uint32_t ref_l1,
|
||||
int ff_hw_base_init_gop_structure(FFHWBaseEncodeContext *ctx, AVCodecContext *avctx,
|
||||
uint32_t ref_l0, uint32_t ref_l1,
|
||||
int flags, int prediction_pre_only)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
|
||||
if (flags & FF_HW_FLAG_INTRA_ONLY || avctx->gop_size <= 1) {
|
||||
av_log(avctx, AV_LOG_VERBOSE, "Using intra frames only.\n");
|
||||
ctx->gop_size = 1;
|
||||
@ -687,9 +680,9 @@ int ff_hw_base_init_gop_structure(AVCodecContext *avctx, uint32_t ref_l0, uint32
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ff_hw_base_get_recon_format(AVCodecContext *avctx, const void *hwconfig, enum AVPixelFormat *fmt)
|
||||
int ff_hw_base_get_recon_format(FFHWBaseEncodeContext *ctx, const void *hwconfig,
|
||||
enum AVPixelFormat *fmt)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
AVHWFramesConstraints *constraints = NULL;
|
||||
enum AVPixelFormat recon_format;
|
||||
int err, i;
|
||||
@ -722,14 +715,14 @@ int ff_hw_base_get_recon_format(AVCodecContext *avctx, const void *hwconfig, enu
|
||||
// No idea what to use; copy input format.
|
||||
recon_format = ctx->input_frames->sw_format;
|
||||
}
|
||||
av_log(avctx, AV_LOG_DEBUG, "Using %s as format of "
|
||||
av_log(ctx->log_ctx, AV_LOG_DEBUG, "Using %s as format of "
|
||||
"reconstructed frames.\n", av_get_pix_fmt_name(recon_format));
|
||||
|
||||
if (ctx->surface_width < constraints->min_width ||
|
||||
ctx->surface_height < constraints->min_height ||
|
||||
ctx->surface_width > constraints->max_width ||
|
||||
ctx->surface_height > constraints->max_height) {
|
||||
av_log(avctx, AV_LOG_ERROR, "Hardware does not support encoding at "
|
||||
av_log(ctx->log_ctx, AV_LOG_ERROR, "Hardware does not support encoding at "
|
||||
"size %dx%d (constraints: width %d-%d height %d-%d).\n",
|
||||
ctx->surface_width, ctx->surface_height,
|
||||
constraints->min_width, constraints->max_width,
|
||||
@ -756,9 +749,9 @@ int ff_hw_base_encode_free(FFHWBaseEncodePicture *pic)
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ff_hw_base_encode_init(AVCodecContext *avctx)
|
||||
int ff_hw_base_encode_init(AVCodecContext *avctx, FFHWBaseEncodeContext *ctx)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
ctx->log_ctx = (void *)avctx;
|
||||
|
||||
ctx->frame = av_frame_alloc();
|
||||
if (!ctx->frame)
|
||||
@ -789,10 +782,8 @@ int ff_hw_base_encode_init(AVCodecContext *avctx)
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ff_hw_base_encode_close(AVCodecContext *avctx)
|
||||
int ff_hw_base_encode_close(FFHWBaseEncodeContext *ctx)
|
||||
{
|
||||
FFHWBaseEncodeContext *ctx = avctx->priv_data;
|
||||
|
||||
av_fifo_freep2(&ctx->encode_fifo);
|
||||
|
||||
av_frame_free(&ctx->frame);
|
||||
|
@ -116,6 +116,7 @@ typedef struct FFHWEncodePictureOperation {
|
||||
|
||||
typedef struct FFHWBaseEncodeContext {
|
||||
const AVClass *class;
|
||||
void *log_ctx;
|
||||
|
||||
// Hardware-specific hooks.
|
||||
const struct FFHWEncodePictureOperation *op;
|
||||
@ -214,21 +215,23 @@ typedef struct FFHWBaseEncodeContext {
|
||||
AVPacket *tail_pkt;
|
||||
} FFHWBaseEncodeContext;
|
||||
|
||||
int ff_hw_base_encode_set_output_property(AVCodecContext *avctx, FFHWBaseEncodePicture *pic,
|
||||
AVPacket *pkt, int flag_no_delay);
|
||||
int ff_hw_base_encode_set_output_property(FFHWBaseEncodeContext *ctx, AVCodecContext *avctx,
|
||||
FFHWBaseEncodePicture *pic, AVPacket *pkt, int flag_no_delay);
|
||||
|
||||
int ff_hw_base_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt);
|
||||
int ff_hw_base_encode_receive_packet(FFHWBaseEncodeContext *ctx, AVCodecContext *avctx, AVPacket *pkt);
|
||||
|
||||
int ff_hw_base_init_gop_structure(AVCodecContext *avctx, uint32_t ref_l0, uint32_t ref_l1,
|
||||
int ff_hw_base_init_gop_structure(FFHWBaseEncodeContext *ctx, AVCodecContext *avctx,
|
||||
uint32_t ref_l0, uint32_t ref_l1,
|
||||
int flags, int prediction_pre_only);
|
||||
|
||||
int ff_hw_base_get_recon_format(AVCodecContext *avctx, const void *hwconfig, enum AVPixelFormat *fmt);
|
||||
int ff_hw_base_get_recon_format(FFHWBaseEncodeContext *ctx, const void *hwconfig,
|
||||
enum AVPixelFormat *fmt);
|
||||
|
||||
int ff_hw_base_encode_free(FFHWBaseEncodePicture *pic);
|
||||
|
||||
int ff_hw_base_encode_init(AVCodecContext *avctx);
|
||||
int ff_hw_base_encode_init(AVCodecContext *avctx, FFHWBaseEncodeContext *ctx);
|
||||
|
||||
int ff_hw_base_encode_close(AVCodecContext *avctx);
|
||||
int ff_hw_base_encode_close(FFHWBaseEncodeContext *ctx);
|
||||
|
||||
#define HW_BASE_ENCODE_COMMON_OPTIONS \
|
||||
{ "idr_interval", \
|
||||
|
@ -811,7 +811,7 @@ static int vaapi_encode_output(AVCodecContext *avctx,
|
||||
av_log(avctx, AV_LOG_DEBUG, "Output read for pic %"PRId64"/%"PRId64".\n",
|
||||
base_pic->display_order, base_pic->encode_order);
|
||||
|
||||
ff_hw_base_encode_set_output_property(avctx, (FFHWBaseEncodePicture*)base_pic, pkt_ptr,
|
||||
ff_hw_base_encode_set_output_property(base_ctx, avctx, (FFHWBaseEncodePicture*)base_pic, pkt_ptr,
|
||||
ctx->codec->flags & FLAG_TIMESTAMP_NO_DELAY);
|
||||
|
||||
end:
|
||||
@ -1658,7 +1658,8 @@ static av_cold int vaapi_encode_init_gop_structure(AVCodecContext *avctx)
|
||||
}
|
||||
#endif
|
||||
|
||||
err = ff_hw_base_init_gop_structure(avctx, ref_l0, ref_l1, ctx->codec->flags, prediction_pre_only);
|
||||
err = ff_hw_base_init_gop_structure(base_ctx, avctx, ref_l0, ref_l1,
|
||||
ctx->codec->flags, prediction_pre_only);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
@ -2059,7 +2060,7 @@ static av_cold int vaapi_encode_create_recon_frames(AVCodecContext *avctx)
|
||||
}
|
||||
hwconfig->config_id = ctx->va_config;
|
||||
|
||||
err = ff_hw_base_get_recon_format(avctx, (const void*)hwconfig, &recon_format);
|
||||
err = ff_hw_base_get_recon_format(base_ctx, (const void*)hwconfig, &recon_format);
|
||||
if (err < 0)
|
||||
goto fail;
|
||||
|
||||
@ -2098,6 +2099,11 @@ static const FFHWEncodePictureOperation vaapi_op = {
|
||||
.free = &vaapi_encode_free,
|
||||
};
|
||||
|
||||
int ff_vaapi_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt)
|
||||
{
|
||||
return ff_hw_base_encode_receive_packet(avctx->priv_data, avctx, pkt);
|
||||
}
|
||||
|
||||
av_cold int ff_vaapi_encode_init(AVCodecContext *avctx)
|
||||
{
|
||||
FFHWBaseEncodeContext *base_ctx = avctx->priv_data;
|
||||
@ -2106,7 +2112,7 @@ av_cold int ff_vaapi_encode_init(AVCodecContext *avctx)
|
||||
VAStatus vas;
|
||||
int err;
|
||||
|
||||
err = ff_hw_base_encode_init(avctx);
|
||||
err = ff_hw_base_encode_init(avctx, base_ctx);
|
||||
if (err < 0)
|
||||
goto fail;
|
||||
|
||||
@ -2313,7 +2319,7 @@ av_cold int ff_vaapi_encode_close(AVCodecContext *avctx)
|
||||
av_freep(&ctx->codec_sequence_params);
|
||||
av_freep(&ctx->codec_picture_params);
|
||||
|
||||
ff_hw_base_encode_close(avctx);
|
||||
ff_hw_base_encode_close(base_ctx);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -343,6 +343,8 @@ typedef struct VAAPIEncodeType {
|
||||
char *data, size_t *data_len);
|
||||
} VAAPIEncodeType;
|
||||
|
||||
int ff_vaapi_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt);
|
||||
|
||||
int ff_vaapi_encode_init(AVCodecContext *avctx);
|
||||
int ff_vaapi_encode_close(AVCodecContext *avctx);
|
||||
|
||||
|
@ -1041,7 +1041,7 @@ const FFCodec ff_av1_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_AV1,
|
||||
.priv_data_size = sizeof(VAAPIEncodeAV1Context),
|
||||
.init = &vaapi_encode_av1_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &vaapi_encode_av1_close,
|
||||
.p.priv_class = &vaapi_encode_av1_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -1385,7 +1385,7 @@ const FFCodec ff_h264_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_H264,
|
||||
.priv_data_size = sizeof(VAAPIEncodeH264Context),
|
||||
.init = &vaapi_encode_h264_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &vaapi_encode_h264_close,
|
||||
.p.priv_class = &vaapi_encode_h264_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -1499,7 +1499,7 @@ const FFCodec ff_hevc_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_HEVC,
|
||||
.priv_data_size = sizeof(VAAPIEncodeH265Context),
|
||||
.init = &vaapi_encode_h265_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &vaapi_encode_h265_close,
|
||||
.p.priv_class = &vaapi_encode_h265_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -574,7 +574,7 @@ const FFCodec ff_mjpeg_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_MJPEG,
|
||||
.priv_data_size = sizeof(VAAPIEncodeMJPEGContext),
|
||||
.init = &vaapi_encode_mjpeg_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &vaapi_encode_mjpeg_close,
|
||||
.p.priv_class = &vaapi_encode_mjpeg_class,
|
||||
.p.capabilities = AV_CODEC_CAP_HARDWARE | AV_CODEC_CAP_DR1 |
|
||||
|
@ -698,7 +698,7 @@ const FFCodec ff_mpeg2_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_MPEG2VIDEO,
|
||||
.priv_data_size = sizeof(VAAPIEncodeMPEG2Context),
|
||||
.init = &vaapi_encode_mpeg2_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &vaapi_encode_mpeg2_close,
|
||||
.p.priv_class = &vaapi_encode_mpeg2_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -252,7 +252,7 @@ const FFCodec ff_vp8_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_VP8,
|
||||
.priv_data_size = sizeof(VAAPIEncodeVP8Context),
|
||||
.init = &vaapi_encode_vp8_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &ff_vaapi_encode_close,
|
||||
.p.priv_class = &vaapi_encode_vp8_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
@ -309,7 +309,7 @@ const FFCodec ff_vp9_vaapi_encoder = {
|
||||
.p.id = AV_CODEC_ID_VP9,
|
||||
.priv_data_size = sizeof(VAAPIEncodeVP9Context),
|
||||
.init = &vaapi_encode_vp9_init,
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_hw_base_encode_receive_packet),
|
||||
FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet),
|
||||
.close = &ff_vaapi_encode_close,
|
||||
.p.priv_class = &vaapi_encode_vp9_class,
|
||||
.p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE |
|
||||
|
Loading…
Reference in New Issue
Block a user