1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-12 19:18:44 +02:00
FFmpeg/libavcodec/vaapi_mpeg2.c

191 lines
7.1 KiB
C
Raw Normal View History

/*
* MPEG-2 HW decode acceleration through VA API
*
* Copyright (C) 2008-2009 Splitted-Desktop Systems
*
* This file is part of Libav.
*
* Libav is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* Libav is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with Libav; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "hwaccel.h"
#include "mpegutils.h"
#include "mpegvideo.h"
#include "internal.h"
#include "vaapi_decode.h"
/** Reconstruct bitstream f_code */
2016-08-06 17:40:14 +02:00
static inline int mpeg2_get_f_code(const MpegEncContext *s)
{
return (s->mpeg_f_code[0][0] << 12) | (s->mpeg_f_code[0][1] << 8) |
(s->mpeg_f_code[1][0] << 4) | s->mpeg_f_code[1][1];
}
/** Determine frame start: first field for field picture or frame picture */
2016-08-06 17:40:14 +02:00
static inline int mpeg2_get_is_frame_start(const MpegEncContext *s)
{
return s->first_field || s->picture_structure == PICT_FRAME;
}
static int vaapi_mpeg2_start_frame(AVCodecContext *avctx, av_unused const uint8_t *buffer, av_unused uint32_t size)
{
2016-08-06 17:40:14 +02:00
const MpegEncContext *s = avctx->priv_data;
VAAPIDecodePicture *pic = s->current_picture_ptr->hwaccel_picture_private;
VAPictureParameterBufferMPEG2 pic_param;
VAIQMatrixBufferMPEG2 iq_matrix;
int i, err;
pic->output_surface = ff_vaapi_get_surface_id(s->current_picture_ptr->f);
pic_param = (VAPictureParameterBufferMPEG2) {
.horizontal_size = s->width,
.vertical_size = s->height,
.forward_reference_picture = VA_INVALID_ID,
.backward_reference_picture = VA_INVALID_ID,
.picture_coding_type = s->pict_type,
.f_code = mpeg2_get_f_code(s),
.picture_coding_extension.bits = {
.intra_dc_precision = s->intra_dc_precision,
.picture_structure = s->picture_structure,
.top_field_first = s->top_field_first,
.frame_pred_frame_dct = s->frame_pred_frame_dct,
.concealment_motion_vectors = s->concealment_motion_vectors,
.q_scale_type = s->q_scale_type,
.intra_vlc_format = s->intra_vlc_format,
.alternate_scan = s->alternate_scan,
.repeat_first_field = s->repeat_first_field,
.progressive_frame = s->progressive_frame,
.is_first_field = mpeg2_get_is_frame_start(s),
},
};
switch (s->pict_type) {
case AV_PICTURE_TYPE_B:
pic_param.backward_reference_picture = ff_vaapi_get_surface_id(s->next_picture.f);
// fall-through
case AV_PICTURE_TYPE_P:
pic_param.forward_reference_picture = ff_vaapi_get_surface_id(s->last_picture.f);
break;
}
err = ff_vaapi_decode_make_param_buffer(avctx, pic,
VAPictureParameterBufferType,
&pic_param, sizeof(pic_param));
if (err < 0)
goto fail;
iq_matrix.load_intra_quantiser_matrix = 1;
iq_matrix.load_non_intra_quantiser_matrix = 1;
iq_matrix.load_chroma_intra_quantiser_matrix = 1;
iq_matrix.load_chroma_non_intra_quantiser_matrix = 1;
for (i = 0; i < 64; i++) {
int n = s->idsp.idct_permutation[ff_zigzag_direct[i]];
iq_matrix.intra_quantiser_matrix[i] = s->intra_matrix[n];
iq_matrix.non_intra_quantiser_matrix[i] = s->inter_matrix[n];
iq_matrix.chroma_intra_quantiser_matrix[i] = s->chroma_intra_matrix[n];
iq_matrix.chroma_non_intra_quantiser_matrix[i] = s->chroma_inter_matrix[n];
}
err = ff_vaapi_decode_make_param_buffer(avctx, pic,
VAIQMatrixBufferType,
&iq_matrix, sizeof(iq_matrix));
if (err < 0)
goto fail;
return 0;
fail:
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
static int vaapi_mpeg2_end_frame(AVCodecContext *avctx)
{
MpegEncContext *s = avctx->priv_data;
VAAPIDecodePicture *pic = s->current_picture_ptr->hwaccel_picture_private;
int ret;
ret = ff_vaapi_decode_issue(avctx, pic);
if (ret < 0)
goto fail;
ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
fail:
return ret;
}
static int vaapi_mpeg2_decode_slice(AVCodecContext *avctx, const uint8_t *buffer, uint32_t size)
{
2016-08-06 17:40:14 +02:00
const MpegEncContext *s = avctx->priv_data;
VAAPIDecodePicture *pic = s->current_picture_ptr->hwaccel_picture_private;
VASliceParameterBufferMPEG2 slice_param;
GetBitContext gb;
uint32_t quantiser_scale_code, intra_slice_flag, macroblock_offset;
int err;
/* Determine macroblock_offset */
init_get_bits(&gb, buffer, 8 * size);
if (get_bits_long(&gb, 32) >> 8 != 1) /* start code */
return AVERROR_INVALIDDATA;
quantiser_scale_code = get_bits(&gb, 5);
intra_slice_flag = get_bits1(&gb);
if (intra_slice_flag) {
skip_bits(&gb, 8);
while (get_bits1(&gb) != 0)
skip_bits(&gb, 8);
}
macroblock_offset = get_bits_count(&gb);
slice_param = (VASliceParameterBufferMPEG2) {
.slice_data_size = size,
.slice_data_offset = 0,
.slice_data_flag = VA_SLICE_DATA_FLAG_ALL,
.macroblock_offset = macroblock_offset,
.slice_horizontal_position = s->mb_x,
.slice_vertical_position = s->mb_y >> (s->picture_structure != PICT_FRAME),
.quantiser_scale_code = quantiser_scale_code,
.intra_slice_flag = intra_slice_flag,
};
err = ff_vaapi_decode_make_slice_buffer(avctx, pic,
&slice_param, sizeof(slice_param),
buffer, size);
if (err < 0) {
ff_vaapi_decode_cancel(avctx, pic);
return err;
}
return 0;
}
const AVHWAccel ff_mpeg2_vaapi_hwaccel = {
.name = "mpeg2_vaapi",
.type = AVMEDIA_TYPE_VIDEO,
.id = AV_CODEC_ID_MPEG2VIDEO,
.pix_fmt = AV_PIX_FMT_VAAPI,
.start_frame = &vaapi_mpeg2_start_frame,
.end_frame = &vaapi_mpeg2_end_frame,
.decode_slice = &vaapi_mpeg2_decode_slice,
.frame_priv_data_size = sizeof(VAAPIDecodePicture),
.init = &ff_vaapi_decode_init,
.uninit = &ff_vaapi_decode_uninit,
.frame_params = &ff_vaapi_common_frame_params,
.priv_data_size = sizeof(VAAPIDecodeContext),
.caps_internal = HWACCEL_CAP_ASYNC_SAFE,
};