1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-12 19:18:44 +02:00
FFmpeg/libavcodec/libxavs2.c

294 lines
10 KiB
C
Raw Normal View History

/*
* AVS2 encoding using the xavs2 library
*
* Copyright (C) 2018 Yiqun Xu, <yiqun.xu@vipl.ict.ac.cn>
* Falei Luo, <falei.luo@gmail.com>
* Huiwen Ren, <hwrenx@gmail.com>
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "xavs2.h"
#include "mpeg12.h"
#include "libavutil/avstring.h"
#define xavs2_opt_set2(name, format, ...) do{ \
char opt_str[16] = {0}; \
int err; \
av_strlcatf(opt_str, sizeof(opt_str), format, __VA_ARGS__); \
err = cae->api->opt_set2(cae->param, name, opt_str); \
if (err < 0) {\
av_log(avctx, AV_LOG_WARNING, "Invalid value for %s: %s\n", name, opt_str);\
}\
} while(0);
typedef struct XAVS2EContext {
AVClass *class;
int lcu_row_threads;
int initial_qp;
int qp;
int max_qp;
int min_qp;
int preset_level;
int log_level;
void *encoder;
char *xavs2_opts;
xavs2_outpacket_t packet;
xavs2_param_t *param;
const xavs2_api_t *api;
} XAVS2EContext;
static av_cold int xavs2_init(AVCodecContext *avctx)
{
XAVS2EContext *cae= avctx->priv_data;
int bit_depth, code;
bit_depth = avctx->pix_fmt == AV_PIX_FMT_YUV420P ? 8 : 10;
/* get API handler */
cae->api = xavs2_api_get(bit_depth);
if (!cae->api) {
av_log(avctx, AV_LOG_ERROR, "api get failed\n");
return AVERROR_EXTERNAL;
}
cae->param = cae->api->opt_alloc();
if (!cae->param) {
av_log(avctx, AV_LOG_ERROR, "param alloc failed\n");
return AVERROR(ENOMEM);
}
xavs2_opt_set2("Width", "%d", avctx->width);
xavs2_opt_set2("Height", "%d", avctx->height);
xavs2_opt_set2("BFrames", "%d", avctx->max_b_frames);
xavs2_opt_set2("BitDepth", "%d", bit_depth);
xavs2_opt_set2("Log", "%d", cae->log_level);
xavs2_opt_set2("Preset", "%d", cae->preset_level);
xavs2_opt_set2("IntraPeriodMax", "%d", avctx->gop_size);
xavs2_opt_set2("IntraPeriodMin", "%d", avctx->gop_size);
xavs2_opt_set2("ThreadFrames", "%d", avctx->thread_count);
xavs2_opt_set2("ThreadRows", "%d", cae->lcu_row_threads);
xavs2_opt_set2("OpenGOP", "%d", !(avctx->flags & AV_CODEC_FLAG_CLOSED_GOP));
if (cae->xavs2_opts) {
AVDictionary *dict = NULL;
AVDictionaryEntry *en = NULL;
if (!av_dict_parse_string(&dict, cae->xavs2_opts, "=", ":", 0)) {
while ((en = av_dict_get(dict, "", en, AV_DICT_IGNORE_SUFFIX))) {
xavs2_opt_set2(en->key, "%s", en->value);
}
av_dict_free(&dict);
}
}
/* Rate control */
if (avctx->bit_rate > 0) {
xavs2_opt_set2("RateControl", "%d", 1);
xavs2_opt_set2("TargetBitRate", "%"PRId64"", avctx->bit_rate);
xavs2_opt_set2("InitialQP", "%d", cae->initial_qp);
xavs2_opt_set2("MaxQP", "%d", avctx->qmax >= 0 ? avctx->qmax : cae->max_qp);
xavs2_opt_set2("MinQP", "%d", avctx->qmin >= 0 ? avctx->qmin : cae->min_qp);
} else {
xavs2_opt_set2("InitialQP", "%d", cae->qp);
}
ff_mpeg12_find_best_frame_rate(avctx->framerate, &code, NULL, NULL, 0);
xavs2_opt_set2("FrameRate", "%d", code);
cae->encoder = cae->api->encoder_create(cae->param);
if (!cae->encoder) {
av_log(avctx,AV_LOG_ERROR, "Can not create encoder. Null pointer returned\n");
return AVERROR(EINVAL);
}
return 0;
}
static void xavs2_copy_frame_with_shift(xavs2_picture_t *pic, const AVFrame *frame, const int shift_in)
{
int j, k;
for (k = 0; k < 3; k++) {
int i_stride = pic->img.i_stride[k];
for (j = 0; j < pic->img.i_lines[k]; j++) {
uint16_t *p_plane = (uint16_t *)&pic->img.img_planes[k][j * i_stride];
int i;
uint8_t *p_buffer = frame->data[k] + frame->linesize[k] * j;
memset(p_plane, 0, i_stride);
for (i = 0; i < pic->img.i_width[k]; i++) {
p_plane[i] = p_buffer[i] << shift_in;
}
}
}
}
static void xavs2_copy_frame(xavs2_picture_t *pic, const AVFrame *frame)
{
int j, k;
for (k = 0; k < 3; k++) {
for (j = 0; j < pic->img.i_lines[k]; j++) {
memcpy( pic->img.img_planes[k] + pic->img.i_stride[k] * j,
frame->data[k]+frame->linesize[k] * j,
pic->img.i_width[k] * pic->img.in_sample_size);
}
}
}
static int xavs2_encode_frame(AVCodecContext *avctx, AVPacket *pkt,
const AVFrame *frame, int *got_packet)
{
XAVS2EContext *cae = avctx->priv_data;
xavs2_picture_t pic;
int ret;
/* create the XAVS2 video encoder */
/* read frame data and send to the XAVS2 video encoder */
if (cae->api->encoder_get_buffer(cae->encoder, &pic) < 0) {
av_log(avctx,AV_LOG_ERROR, "failed to get frame buffer\n");
return AVERROR_EXTERNAL;
}
if (frame) {
switch (frame->format) {
case AV_PIX_FMT_YUV420P:
if (pic.img.in_sample_size == pic.img.enc_sample_size) {
xavs2_copy_frame(&pic, frame);
} else {
const int shift_in = atoi(cae->api->opt_get(cae->param, "SampleShift"));
xavs2_copy_frame_with_shift(&pic, frame, shift_in);
}
break;
case AV_PIX_FMT_YUV420P10:
if (pic.img.in_sample_size == pic.img.enc_sample_size) {
xavs2_copy_frame(&pic, frame);
break;
}
default:
av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format\n");
return AVERROR(EINVAL);
break;
}
pic.i_state = 0;
pic.i_pts = frame->pts;
pic.i_type = XAVS2_TYPE_AUTO;
ret = cae->api->encoder_encode(cae->encoder, &pic, &cae->packet);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "encode failed\n");
return AVERROR_EXTERNAL;
}
} else {
cae->api->encoder_encode(cae->encoder, NULL, &cae->packet);
}
if ((cae->packet.len) && (cae->packet.state != XAVS2_STATE_FLUSH_END)){
if (av_new_packet(pkt, cae->packet.len) < 0){
av_log(avctx, AV_LOG_ERROR, "packet alloc failed\n");
cae->api->encoder_packet_unref(cae->encoder, &cae->packet);
return AVERROR(ENOMEM);
}
pkt->pts = cae->packet.pts;
pkt->dts = cae->packet.dts;
memcpy(pkt->data, cae->packet.stream, cae->packet.len);
pkt->size = cae->packet.len;
cae->api->encoder_packet_unref(cae->encoder, &cae->packet);
*got_packet = 1;
} else {
*got_packet = 0;
}
return 0;
}
static av_cold int xavs2_close(AVCodecContext *avctx)
{
XAVS2EContext *cae = avctx->priv_data;
/* destroy the encoder */
if (cae->api) {
cae->api->encoder_destroy(cae->encoder);
if (cae->param) {
cae->api->opt_destroy(cae->param);
}
}
return 0;
}
#define OFFSET(x) offsetof(XAVS2EContext, x)
#define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
static const AVOption options[] = {
{ "lcu_row_threads" , "number of parallel threads for rows" , OFFSET(lcu_row_threads) , AV_OPT_TYPE_INT, {.i64 = 0 }, 0, INT_MAX, VE },
{ "initial_qp" , "Quantization initial parameter" , OFFSET(initial_qp) , AV_OPT_TYPE_INT, {.i64 = 34 }, 1, 63, VE },
{ "qp" , "Quantization parameter" , OFFSET(qp) , AV_OPT_TYPE_INT, {.i64 = 34 }, 1, 63, VE },
{ "max_qp" , "max qp for rate control" , OFFSET(max_qp) , AV_OPT_TYPE_INT, {.i64 = 55 }, 0, 63, VE },
{ "min_qp" , "min qp for rate control" , OFFSET(min_qp) , AV_OPT_TYPE_INT, {.i64 = 20 }, 0, 63, VE },
{ "speed_level" , "Speed level, higher is better but slower", OFFSET(preset_level) , AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 9, VE },
{ "log_level" , "log level: -1: none, 0: error, 1: warning, 2: info, 3: debug", OFFSET(log_level) , AV_OPT_TYPE_INT, {.i64 = 0 }, -1, 3, VE },
{ "xavs2-params" , "set the xavs2 configuration using a :-separated list of key=value parameters", OFFSET(xavs2_opts), AV_OPT_TYPE_STRING, { 0 }, 0, 0, VE },
{ NULL },
};
static const AVClass libxavs2 = {
.class_name = "XAVS2EContext",
.item_name = av_default_item_name,
.option = options,
.version = LIBAVUTIL_VERSION_INT,
};
static const AVCodecDefault xavs2_defaults[] = {
{ "b", "0" },
{ "g", "48"},
{ "bf", "7" },
{ NULL },
};
AVCodec ff_libxavs2_encoder = {
.name = "libxavs2",
.long_name = NULL_IF_CONFIG_SMALL("libxavs2 AVS2-P2/IEEE1857.4"),
.type = AVMEDIA_TYPE_VIDEO,
.id = AV_CODEC_ID_AVS2,
.priv_data_size = sizeof(XAVS2EContext),
.init = xavs2_init,
.encode2 = xavs2_encode_frame,
.close = xavs2_close,
.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_AUTO_THREADS,
.pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUV420P,
AV_PIX_FMT_NONE },
.priv_class = &libxavs2,
.defaults = xavs2_defaults,
.wrapper_name = "libxavs2",
} ;