1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2025-08-04 22:03:09 +02:00

avcodec/mpegvideoenc: Move ratecontrol to MPVMainEncContext

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
This commit is contained in:
Andreas Rheinhardt
2022-01-28 20:43:54 +01:00
parent bb0a72b100
commit d8c7925e32
6 changed files with 140 additions and 120 deletions

View File

@ -42,7 +42,6 @@
#include "mpegvideoencdsp.h"
#include "pixblockdsp.h"
#include "put_bits.h"
#include "ratecontrol.h"
#include "qpeldsp.h"
#include "videodsp.h"
@ -327,13 +326,6 @@ typedef struct MpegEncContext {
int dct_count[2];
uint16_t (*dct_offset)[64];
/* bit rate control */
int64_t total_bits;
int frame_bits; ///< bits used for the current frame
int stuffing_bits; ///< bits used for stuffing
int next_lambda; ///< next lambda used for retrying to encode a frame
RateControlContext rc_context; ///< contains stuff only accessed in ratecontrol.c
/* statistics, used for 2-pass encoding */
int mv_bits;
int header_bits;

View File

@ -84,7 +84,7 @@
#define QMAT_SHIFT_MMX 16
#define QMAT_SHIFT 21
static int encode_picture(MpegEncContext *s, const AVPacket *pkt);
static int encode_picture(MPVMainEncContext *const s, const AVPacket *pkt);
static int dct_quantize_refine(MpegEncContext *s, int16_t *block, int16_t *weight, int16_t *orig, int n, int qscale);
static int sse_mb(MpegEncContext *s);
static void denoise_dct_c(MpegEncContext *s, int16_t *block);
@ -1008,7 +1008,8 @@ av_cold int ff_mpv_encode_init(AVCodecContext *avctx)
#endif
}
if ((ret = ff_rate_control_init(s)) < 0)
ret = ff_rate_control_init(m);
if (ret < 0)
return ret;
if (m->b_frame_strategy == 2) {
@ -1043,7 +1044,7 @@ av_cold int ff_mpv_encode_end(AVCodecContext *avctx)
MPVMainEncContext *const m = avctx->priv_data;
MpegEncContext *const s = &m->s;
ff_rate_control_uninit(&s->rc_context);
ff_rate_control_uninit(&m->rc_context);
ff_mpv_common_end(s);
av_refstruct_pool_uninit(&s->picture_pool);
@ -1622,7 +1623,7 @@ static int set_bframe_chain_length(MPVMainEncContext *const m)
// FIXME check that the gop check above is +-1 correct
av_refstruct_unref(&s->input_picture[0]);
ff_vbv_update(s, 0);
ff_vbv_update(m, 0);
return 0;
}
@ -1642,15 +1643,15 @@ static int set_bframe_chain_length(MPVMainEncContext *const m)
for (int i = 0; i < s->max_b_frames + 1; i++) {
int pict_num = s->input_picture[0]->display_picture_number + i;
if (pict_num >= s->rc_context.num_entries)
if (pict_num >= m->rc_context.num_entries)
break;
if (!s->input_picture[i]) {
s->rc_context.entry[pict_num - 1].new_pict_type = AV_PICTURE_TYPE_P;
m->rc_context.entry[pict_num - 1].new_pict_type = AV_PICTURE_TYPE_P;
break;
}
s->input_picture[i]->f->pict_type =
s->rc_context.entry[pict_num].new_pict_type;
m->rc_context.entry[pict_num].new_pict_type;
}
}
@ -1918,7 +1919,7 @@ int ff_mpv_encode_picture(AVCodecContext *avctx, AVPacket *pkt,
//emms_c();
frame_start(s);
vbv_retry:
ret = encode_picture(s, pkt);
ret = encode_picture(m, pkt);
if (growing_buffer) {
av_assert0(s->pb.buf == avctx->internal->byte_buffer);
pkt->data = s->pb.buf;
@ -1933,14 +1934,14 @@ vbv_retry:
ff_mjpeg_encode_picture_trailer(&s->pb, s->header_bits);
if (avctx->rc_buffer_size) {
RateControlContext *rcc = &s->rc_context;
RateControlContext *rcc = &m->rc_context;
int max_size = FFMAX(rcc->buffer_index * avctx->rc_max_available_vbv_use, rcc->buffer_index - 500);
int hq = (avctx->mb_decision == FF_MB_DECISION_RD || avctx->trellis);
int min_step = hq ? 1 : (1<<(FF_LAMBDA_SHIFT + 7))/139;
if (put_bits_count(&s->pb) > max_size &&
s->lambda < s->lmax) {
s->next_lambda = FFMAX(s->lambda + min_step, s->lambda *
m->next_lambda = FFMAX(s->lambda + min_step, s->lambda *
(s->qscale + 1) / s->qscale);
if (s->adaptive_quant) {
int i;
@ -1968,7 +1969,7 @@ vbv_retry:
}
if (avctx->flags & AV_CODEC_FLAG_PASS1)
ff_write_pass1_stats(s);
ff_write_pass1_stats(m);
for (int i = 0; i < MPV_MAX_PLANES; i++)
avctx->error[i] += s->encoding_error[i];
@ -1982,10 +1983,10 @@ vbv_retry:
s->misc_bits + s->i_tex_bits +
s->p_tex_bits);
flush_put_bits(&s->pb);
s->frame_bits = put_bits_count(&s->pb);
m->frame_bits = put_bits_count(&s->pb);
stuffing_count = ff_vbv_update(s, s->frame_bits);
s->stuffing_bits = 8*stuffing_count;
stuffing_count = ff_vbv_update(m, m->frame_bits);
m->stuffing_bits = 8*stuffing_count;
if (stuffing_count) {
if (put_bytes_left(&s->pb, 0) < stuffing_count + 50) {
av_log(avctx, AV_LOG_ERROR, "stuffing too large\n");
@ -2009,10 +2010,10 @@ vbv_retry:
break;
default:
av_log(avctx, AV_LOG_ERROR, "vbv buffer overflow\n");
s->stuffing_bits = 0;
m->stuffing_bits = 0;
}
flush_put_bits(&s->pb);
s->frame_bits = put_bits_count(&s->pb);
m->frame_bits = put_bits_count(&s->pb);
}
/* update MPEG-1/2 vbv_delay for CBR */
@ -2027,9 +2028,9 @@ vbv_retry:
int vbv_delay, min_delay;
double inbits = avctx->rc_max_rate *
av_q2d(avctx->time_base);
int minbits = s->frame_bits - 8 *
int minbits = m->frame_bits - 8 *
(s->vbv_delay_pos - 1);
double bits = s->rc_context.buffer_index + minbits - inbits;
double bits = m->rc_context.buffer_index + minbits - inbits;
uint8_t *const vbv_delay_ptr = s->pb.buf + s->vbv_delay_pos;
if (bits < 0)
@ -2064,7 +2065,7 @@ vbv_retry:
return ret;
}
}
s->total_bits += s->frame_bits;
m->total_bits += m->frame_bits;
pkt->pts = s->cur_pic.ptr->f->pts;
pkt->duration = s->cur_pic.ptr->f->duration;
@ -2089,14 +2090,14 @@ vbv_retry:
if (s->mb_info)
av_packet_shrink_side_data(pkt, AV_PKT_DATA_H263_MB_INFO, s->mb_info_size);
} else {
s->frame_bits = 0;
m->frame_bits = 0;
}
ff_mpv_unref_picture(&s->cur_pic);
av_assert1((s->frame_bits & 7) == 0);
av_assert1((m->frame_bits & 7) == 0);
pkt->size = s->frame_bits / 8;
pkt->size = m->frame_bits / 8;
*got_packet = !!pkt->size;
return 0;
}
@ -3611,12 +3612,15 @@ static void merge_context_after_encode(MpegEncContext *dst, MpegEncContext *src)
flush_put_bits(&dst->pb);
}
static int estimate_qp(MpegEncContext *s, int dry_run){
if (s->next_lambda){
s->cur_pic.ptr->f->quality = s->next_lambda;
if(!dry_run) s->next_lambda= 0;
static int estimate_qp(MPVMainEncContext *const m, int dry_run)
{
MpegEncContext *const s = &m->s;
if (m->next_lambda){
s->cur_pic.ptr->f->quality = m->next_lambda;
if(!dry_run) m->next_lambda= 0;
} else if (!s->fixed_qscale) {
int quality = ff_rate_estimate_qscale(s, dry_run);
int quality = ff_rate_estimate_qscale(m, dry_run);
s->cur_pic.ptr->f->quality = quality;
if (s->cur_pic.ptr->f->quality < 0)
return -1;
@ -3661,8 +3665,9 @@ static void set_frame_distances(MpegEncContext * s){
}
}
static int encode_picture(MpegEncContext *s, const AVPacket *pkt)
static int encode_picture(MPVMainEncContext *const m, const AVPacket *pkt)
{
MpegEncContext *const s = &m->s;
int i, ret;
int bits;
int context_count = s->slice_context_count;
@ -3689,9 +3694,10 @@ static int encode_picture(MpegEncContext *s, const AVPacket *pkt)
}
if (s->avctx->flags & AV_CODEC_FLAG_PASS2) {
if (estimate_qp(s,1) < 0)
return -1;
ff_get_2pass_fcode(s);
ret = estimate_qp(m, 1);
if (ret < 0)
return ret;
ff_get_2pass_fcode(m);
} else if (!(s->avctx->flags & AV_CODEC_FLAG_QSCALE)) {
if(s->pict_type==AV_PICTURE_TYPE_B)
s->lambda= s->last_lambda_for[s->pict_type];
@ -3814,8 +3820,9 @@ static int encode_picture(MpegEncContext *s, const AVPacket *pkt)
}
}
if (estimate_qp(s, 0) < 0)
return -1;
ret = estimate_qp(m, 0);
if (ret < 0)
return ret;
if (s->qscale < 3 && s->max_qcoeff <= 128 &&
s->pict_type == AV_PICTURE_TYPE_I &&

View File

@ -32,6 +32,7 @@
#include "libavutil/opt.h"
#include "mpegvideo.h"
#include "ratecontrol.h"
#define MPVENC_MAX_B_FRAMES 16
@ -43,6 +44,13 @@ typedef struct MPVMainEncContext {
int b_frame_strategy;
int b_sensitivity;
int brd_scale;
/* bit rate control */
int64_t total_bits;
int frame_bits; ///< bits used for the current frame
int stuffing_bits; ///< bits used for stuffing
int next_lambda; ///< next lambda used for retrying to encode a frame
RateControlContext rc_context; ///< contains stuff only accessed in ratecontrol.c
} MPVMainEncContext;
#define MAX_FCODE 7
@ -94,7 +102,7 @@ typedef struct MPVMainEncContext {
#define FF_MPV_OFFSET(x) offsetof(MpegEncContext, x)
#define FF_MPV_MAIN_OFFSET(x) offsetof(MPVMainEncContext, x)
#define FF_RC_OFFSET(x) offsetof(MpegEncContext, rc_context.x)
#define FF_RC_OFFSET(x) offsetof(MPVMainEncContext, rc_context.x)
#define FF_MPV_OPT_FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM)
#define FF_MPV_COMMON_OPTS \
FF_MPV_OPT_CMP_FUNC, \

View File

@ -35,8 +35,9 @@
#include "mpegvideoenc.h"
#include "libavutil/eval.h"
void ff_write_pass1_stats(MpegEncContext *s)
void ff_write_pass1_stats(MPVMainEncContext *const m)
{
const MpegEncContext *const s = &m->s;
snprintf(s->avctx->stats_out, 256,
"in:%d out:%d type:%d q:%d itex:%d ptex:%d mv:%d misc:%d "
"fcode:%d bcode:%d mc-var:%"PRId64" var:%"PRId64" icount:%d hbits:%d;\n",
@ -101,9 +102,10 @@ static double bits2qp_cb(void *rce, double qp)
return bits2qp(rce, qp);
}
static double get_diff_limited_q(MpegEncContext *s, const RateControlEntry *rce, double q)
static double get_diff_limited_q(MPVMainEncContext *m, const RateControlEntry *rce, double q)
{
RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
RateControlContext *const rcc = &m->rc_context;
AVCodecContext *a = s->avctx;
const int pict_type = rce->new_pict_type;
const double last_p_q = rcc->last_qscale_for[AV_PICTURE_TYPE_P];
@ -168,10 +170,11 @@ static void get_qminmax(int *qmin_ret, int *qmax_ret, MpegEncContext *s, int pic
*qmax_ret = qmax;
}
static double modify_qscale(MpegEncContext *s, const RateControlEntry *rce,
static double modify_qscale(MPVMainEncContext *const m, const RateControlEntry *rce,
double q, int frame_num)
{
RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
RateControlContext *const rcc = &m->rc_context;
const double buffer_size = s->avctx->rc_buffer_size;
const double fps = get_fps(s->avctx);
const double min_rate = s->avctx->rc_min_rate / fps;
@ -259,10 +262,11 @@ static double modify_qscale(MpegEncContext *s, const RateControlEntry *rce,
/**
* Modify the bitrate curve from pass1 for one frame.
*/
static double get_qscale(MpegEncContext *s, RateControlEntry *rce,
static double get_qscale(MPVMainEncContext *const m, RateControlEntry *rce,
double rate_factor, int frame_num)
{
RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
RateControlContext *rcc = &m->rc_context;
AVCodecContext *a = s->avctx;
const int pict_type = rce->new_pict_type;
const double mb_num = s->mb_num;
@ -332,9 +336,10 @@ static double get_qscale(MpegEncContext *s, RateControlEntry *rce,
return q;
}
static int init_pass2(MpegEncContext *s)
static int init_pass2(MPVMainEncContext *const m)
{
RateControlContext *rcc = &s->rc_context;
RateControlContext *const rcc = &m->rc_context;
MpegEncContext *const s = &m->s;
AVCodecContext *a = s->avctx;
int i, toobig;
AVRational fps = get_fpsQ(s->avctx);
@ -393,7 +398,7 @@ static int init_pass2(MpegEncContext *s)
for (i = 0; i < rcc->num_entries; i++) {
const RateControlEntry *rce = &rcc->entry[i];
qscale[i] = get_qscale(s, &rcc->entry[i], rate_factor, i);
qscale[i] = get_qscale(m, &rcc->entry[i], rate_factor, i);
rcc->last_qscale_for[rce->pict_type] = qscale[i];
}
av_assert0(filter_size % 2 == 1);
@ -402,13 +407,13 @@ static int init_pass2(MpegEncContext *s)
for (i = FFMAX(0, rcc->num_entries - 300); i < rcc->num_entries; i++) {
const RateControlEntry *rce = &rcc->entry[i];
qscale[i] = get_diff_limited_q(s, rce, qscale[i]);
qscale[i] = get_diff_limited_q(m, rce, qscale[i]);
}
for (i = rcc->num_entries - 1; i >= 0; i--) {
const RateControlEntry *rce = &rcc->entry[i];
qscale[i] = get_diff_limited_q(s, rce, qscale[i]);
qscale[i] = get_diff_limited_q(m, rce, qscale[i]);
}
/* smooth curve */
@ -438,10 +443,10 @@ static int init_pass2(MpegEncContext *s)
RateControlEntry *rce = &rcc->entry[i];
double bits;
rce->new_qscale = modify_qscale(s, rce, blurred_qscale[i], i);
rce->new_qscale = modify_qscale(m, rce, blurred_qscale[i], i);
bits = qp2bits(rce, rce->new_qscale) + rce->mv_bits + rce->misc_bits;
bits += 8 * ff_vbv_update(s, bits);
bits += 8 * ff_vbv_update(m, bits);
rce->expected_bits = expected_bits;
expected_bits += bits;
@ -494,9 +499,10 @@ static int init_pass2(MpegEncContext *s)
return 0;
}
av_cold int ff_rate_control_init(MpegEncContext *s)
av_cold int ff_rate_control_init(MPVMainEncContext *const m)
{
RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
RateControlContext *rcc = &m->rc_context;
int i, res;
static const char * const const_names[] = {
"PI",
@ -632,7 +638,7 @@ av_cold int ff_rate_control_init(MpegEncContext *s)
p = next;
}
res = init_pass2(s);
res = init_pass2(m);
if (res < 0)
return res;
}
@ -686,7 +692,7 @@ av_cold int ff_rate_control_init(MpegEncContext *s)
rcc->mv_bits_sum[rce.pict_type] += rce.mv_bits;
rcc->frame_count[rce.pict_type]++;
get_qscale(s, &rce, rcc->pass1_wanted_bits / rcc->pass1_rc_eq_output_sum, i);
get_qscale(m, &rce, rcc->pass1_wanted_bits / rcc->pass1_rc_eq_output_sum, i);
// FIXME misbehaves a little for variable fps
rcc->pass1_wanted_bits += s->bit_rate / get_fps(s->avctx);
@ -717,9 +723,10 @@ av_cold void ff_rate_control_uninit(RateControlContext *rcc)
av_freep(&rcc->cplx_tab);
}
int ff_vbv_update(MpegEncContext *s, int frame_size)
int ff_vbv_update(MPVMainEncContext *m, int frame_size)
{
RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
RateControlContext *const rcc = &m->rc_context;
const double fps = get_fps(s->avctx);
const int buffer_size = s->avctx->rc_buffer_size;
const double min_rate = s->avctx->rc_min_rate / fps;
@ -893,9 +900,10 @@ static void adaptive_quantization(RateControlContext *const rcc,
}
}
void ff_get_2pass_fcode(MpegEncContext *s)
void ff_get_2pass_fcode(MPVMainEncContext *const m)
{
const RateControlContext *rcc = &s->rc_context;
MpegEncContext *const s = &m->s;
const RateControlContext *rcc = &m->rc_context;
const RateControlEntry *rce = &rcc->entry[s->picture_number];
s->f_code = rce->f_code;
@ -904,8 +912,10 @@ void ff_get_2pass_fcode(MpegEncContext *s)
// FIXME rd or at least approx for dquant
float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
float ff_rate_estimate_qscale(MPVMainEncContext *const m, int dry_run)
{
MpegEncContext *const s = &m->s;
RateControlContext *rcc = &m->rc_context;
float q;
int qmin, qmax;
float br_compensation;
@ -914,7 +924,6 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
double fps;
int picture_number = s->picture_number;
int64_t wanted_bits;
RateControlContext *rcc = &s->rc_context;
AVCodecContext *a = s->avctx;
RateControlEntry local_rce, *rce;
double bits;
@ -931,11 +940,11 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
const int64_t last_var =
s->last_pict_type == AV_PICTURE_TYPE_I ? rcc->last_mb_var_sum
: rcc->last_mc_mb_var_sum;
av_assert1(s->frame_bits >= s->stuffing_bits);
av_assert1(m->frame_bits >= m->stuffing_bits);
update_predictor(&rcc->pred[s->last_pict_type],
rcc->last_qscale,
sqrt(last_var),
s->frame_bits - s->stuffing_bits);
m->frame_bits - m->stuffing_bits);
}
if (s->avctx->flags & AV_CODEC_FLAG_PASS2) {
@ -970,7 +979,7 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
wanted_bits = (int64_t)wanted_bits_double;
}
diff = s->total_bits - wanted_bits;
diff = m->total_bits - wanted_bits;
br_compensation = (a->bit_rate_tolerance - diff) / a->bit_rate_tolerance;
if (br_compensation <= 0.0)
br_compensation = 0.001;
@ -984,7 +993,7 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
q = rce->new_qscale / br_compensation;
ff_dlog(s->avctx, "%f %f %f last:%d var:%"PRId64" type:%d//\n", q, rce->new_qscale,
br_compensation, s->frame_bits, var, pict_type);
br_compensation, m->frame_bits, var, pict_type);
} else {
rce->pict_type =
rce->new_pict_type = pict_type;
@ -1015,12 +1024,12 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
rate_factor = rcc->pass1_wanted_bits /
rcc->pass1_rc_eq_output_sum * br_compensation;
q = get_qscale(s, rce, rate_factor, picture_number);
q = get_qscale(m, rce, rate_factor, picture_number);
if (q < 0)
return -1;
av_assert0(q > 0.0);
q = get_diff_limited_q(s, rce, q);
q = get_diff_limited_q(m, rce, q);
av_assert0(q > 0.0);
// FIXME type dependent blur like in 2-pass
@ -1034,7 +1043,7 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
}
av_assert0(q > 0.0);
q = modify_qscale(s, rce, q, picture_number);
q = modify_qscale(m, rce, q, picture_number);
rcc->pass1_wanted_bits += s->bit_rate / fps;
@ -1047,8 +1056,8 @@ float ff_rate_estimate_qscale(MpegEncContext *s, int dry_run)
"size:%d var:%"PRId64"/%"PRId64" br:%"PRId64" fps:%d\n",
av_get_picture_type_char(pict_type),
qmin, q, qmax, picture_number,
wanted_bits / 1000, s->total_bits / 1000,
br_compensation, short_term_q, s->frame_bits,
wanted_bits / 1000, m->total_bits / 1000,
br_compensation, short_term_q, m->frame_bits,
s->mb_var_sum, s->mc_mb_var_sum,
s->bit_rate / 1000, (int)fps);
}

View File

@ -93,14 +93,14 @@ typedef struct RateControlContext{
float *cplx_tab, *bits_tab;
}RateControlContext;
struct MpegEncContext;
typedef struct MPVMainEncContext MPVMainEncContext;
/* rate control */
int ff_rate_control_init(struct MpegEncContext *s);
float ff_rate_estimate_qscale(struct MpegEncContext *s, int dry_run);
void ff_write_pass1_stats(struct MpegEncContext *s);
int ff_vbv_update(struct MpegEncContext *s, int frame_size);
void ff_get_2pass_fcode(struct MpegEncContext *s);
int ff_rate_control_init(MPVMainEncContext *m);
float ff_rate_estimate_qscale(MPVMainEncContext *m, int dry_run);
void ff_write_pass1_stats(MPVMainEncContext *m);
int ff_vbv_update(MPVMainEncContext *m, int frame_size);
void ff_get_2pass_fcode(MPVMainEncContext *m);
void ff_rate_control_uninit(RateControlContext *rcc);
#endif /* AVCODEC_RATECONTROL_H */

View File

@ -61,7 +61,7 @@ typedef struct SnowEncContext {
int scenechange_threshold;
MECmpContext mecc;
MpegEncContext m; // needed for motion estimation, should not be used for anything else, the idea is to eventually make the motion estimation independent of MpegEncContext, so this will be removed then (FIXME/XXX)
MPVMainEncContext m; // needed for motion estimation, should not be used for anything else, the idea is to eventually make the motion estimation independent of MpegEncContext, so this will be removed then (FIXME/XXX)
MPVPicture cur_pic, last_pic;
#define ME_CACHE_SIZE 1024
unsigned me_cache[ME_CACHE_SIZE];
@ -160,7 +160,7 @@ static av_cold int encode_init(AVCodecContext *avctx)
{
SnowEncContext *const enc = avctx->priv_data;
SnowContext *const s = &enc->com;
MpegEncContext *const mpv = &enc->m;
MpegEncContext *const mpv = &enc->m.s;
int plane_index, ret;
int i;
@ -217,7 +217,7 @@ static av_cold int encode_init(AVCodecContext *avctx)
mcf(12,12)
ff_me_cmp_init(&enc->mecc, avctx);
ret = ff_me_init(&enc->m.me, avctx, &enc->mecc, 0);
ret = ff_me_init(&mpv->me, avctx, &enc->mecc, 0);
if (ret < 0)
return ret;
ff_mpegvideoencdsp_init(&enc->mpvencdsp, avctx);
@ -252,7 +252,7 @@ static av_cold int encode_init(AVCodecContext *avctx)
return AVERROR(ENOMEM);
}
if((avctx->flags&AV_CODEC_FLAG_PASS2) || !(avctx->flags&AV_CODEC_FLAG_QSCALE)){
ret = ff_rate_control_init(mpv);
ret = ff_rate_control_init(&enc->m);
if(ret < 0)
return ret;
}
@ -369,7 +369,7 @@ static inline int get_penalty_factor(int lambda, int lambda2, int type){
static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
{
SnowContext *const s = &enc->com;
MotionEstContext *const c = &enc->m.me;
MotionEstContext *const c = &enc->m.s.me;
uint8_t p_buffer[1024];
uint8_t i_buffer[1024];
uint8_t p_state[sizeof(s->block_state)];
@ -435,9 +435,9 @@ static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
last_mv[2][0]= bottom->mx;
last_mv[2][1]= bottom->my;
enc->m.mb_stride = 2;
enc->m.mb_x =
enc->m.mb_y = 0;
enc->m.s.mb_stride = 2;
enc->m.s.mb_x =
enc->m.s.mb_y = 0;
c->skip= 0;
av_assert1(c-> stride == stride);
@ -446,7 +446,7 @@ static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
c->penalty_factor = get_penalty_factor(enc->lambda, enc->lambda2, c->avctx->me_cmp);
c->sub_penalty_factor= get_penalty_factor(enc->lambda, enc->lambda2, c->avctx->me_sub_cmp);
c->mb_penalty_factor = get_penalty_factor(enc->lambda, enc->lambda2, c->avctx->mb_cmp);
c->current_mv_penalty = c->mv_penalty[enc->m.f_code=1] + MAX_DMV;
c->current_mv_penalty = c->mv_penalty[enc->m.s.f_code=1] + MAX_DMV;
c->xmin = - x*block_w - 16+3;
c->ymin = - y*block_w - 16+3;
@ -482,7 +482,7 @@ static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
for(ref=0; ref<s->ref_frames; ref++){
init_ref(c, current_data, s->last_picture[ref]->data, NULL, block_w*x, block_w*y, 0);
ref_score= ff_epzs_motion_search(&enc->m, &ref_mx, &ref_my, P, 0, /*ref_index*/ 0, last_mv,
ref_score = ff_epzs_motion_search(&enc->m.s, &ref_mx, &ref_my, P, 0, /*ref_index*/ 0, last_mv,
(1<<16)>>shift, level-LOG2_MB_SIZE+4, block_w);
av_assert2(ref_mx >= c->xmin);
@ -490,8 +490,10 @@ static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
av_assert2(ref_my >= c->ymin);
av_assert2(ref_my <= c->ymax);
ref_score= c->sub_motion_search(&enc->m, &ref_mx, &ref_my, ref_score, 0, 0, level-LOG2_MB_SIZE+4, block_w);
ref_score= ff_get_mb_score(&enc->m, ref_mx, ref_my, 0, 0, level-LOG2_MB_SIZE+4, block_w, 0);
ref_score = c->sub_motion_search(&enc->m.s, &ref_mx, &ref_my, ref_score,
0, 0, level-LOG2_MB_SIZE+4, block_w);
ref_score = ff_get_mb_score(&enc->m.s, ref_mx, ref_my, 0, 0,
level-LOG2_MB_SIZE+4, block_w, 0);
ref_score+= 2*av_log2(2*ref)*c->penalty_factor;
if(s->ref_mvs[ref]){
s->ref_mvs[ref][index][0]= ref_mx;
@ -567,7 +569,7 @@ static int encode_q_branch(SnowEncContext *enc, int level, int x, int y)
if (vard <= 64 || vard < varc)
c->scene_change_score+= ff_sqrt(vard) - ff_sqrt(varc);
else
c->scene_change_score += enc->m.qscale;
c->scene_change_score += enc->m.s.qscale;
}
if(level!=s->block_max_depth){
@ -670,7 +672,7 @@ static int get_dc(SnowEncContext *enc, int mb_x, int mb_y, int plane_index)
const int obmc_stride= plane_index ? (2*block_size)>>s->chroma_h_shift : 2*block_size;
const int ref_stride= s->current_picture->linesize[plane_index];
const uint8_t *src = s->input_picture->data[plane_index];
IDWTELEM *dst= (IDWTELEM*)enc->m.sc.obmc_scratchpad + plane_index*block_size*block_size*4; //FIXME change to unsigned
IDWTELEM *dst= (IDWTELEM*)enc->m.s.sc.obmc_scratchpad + plane_index*block_size*block_size*4; //FIXME change to unsigned
const int b_stride = s->b_width << s->block_max_depth;
const int w= p->width;
const int h= p->height;
@ -768,7 +770,7 @@ static int get_block_rd(SnowEncContext *enc, int mb_x, int mb_y,
const int ref_stride= s->current_picture->linesize[plane_index];
uint8_t *dst= s->current_picture->data[plane_index];
const uint8_t *src = s->input_picture->data[plane_index];
IDWTELEM *pred= (IDWTELEM*)enc->m.sc.obmc_scratchpad + plane_index*block_size*block_size*4;
IDWTELEM *pred= (IDWTELEM*)enc->m.s.sc.obmc_scratchpad + plane_index*block_size*block_size*4;
uint8_t *cur = s->scratchbuf;
uint8_t *tmp = s->emu_edge_buffer;
const int b_stride = s->b_width << s->block_max_depth;
@ -831,19 +833,19 @@ static int get_block_rd(SnowEncContext *enc, int mb_x, int mb_y,
* to improve the score of the whole frame, thus iterative motion
* estimation does not always converge. */
if(s->avctx->me_cmp == FF_CMP_W97)
distortion = ff_w97_32_c(&enc->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32);
distortion = ff_w97_32_c(&enc->m.s, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32);
else if(s->avctx->me_cmp == FF_CMP_W53)
distortion = ff_w53_32_c(&enc->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32);
distortion = ff_w53_32_c(&enc->m.s, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32);
else{
distortion = 0;
for(i=0; i<4; i++){
int off = sx+16*(i&1) + (sy+16*(i>>1))*ref_stride;
distortion += enc->m.me.me_cmp[0](&enc->m, src + off, dst + off, ref_stride, 16);
distortion += enc->m.s.me.me_cmp[0](&enc->m.s, src + off, dst + off, ref_stride, 16);
}
}
}else{
av_assert2(block_w==8);
distortion = enc->m.me.me_cmp[0](&enc->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, block_w*2);
distortion = enc->m.s.me.me_cmp[0](&enc->m.s, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, block_w*2);
}
if(plane_index==0){
@ -909,7 +911,7 @@ static int get_4block_rd(SnowEncContext *enc, int mb_x, int mb_y, int plane_inde
}
av_assert1(block_w== 8 || block_w==16);
distortion += enc->m.me.me_cmp[block_w==8](&enc->m, src + x + y*ref_stride, dst + x + y*ref_stride, ref_stride, block_h);
distortion += enc->m.s.me.me_cmp[block_w==8](&enc->m.s, src + x + y*ref_stride, dst + x + y*ref_stride, ref_stride, block_h);
}
if(plane_index==0){
@ -1706,11 +1708,11 @@ static int ratecontrol_1pass(SnowEncContext *enc, AVFrame *pict)
coef_sum = (uint64_t)coef_sum * coef_sum >> 16;
if(pict->pict_type == AV_PICTURE_TYPE_I){
enc->m.mb_var_sum = coef_sum;
enc->m.mc_mb_var_sum = 0;
enc->m.s.mb_var_sum = coef_sum;
enc->m.s.mc_mb_var_sum = 0;
}else{
enc->m.mc_mb_var_sum = coef_sum;
enc->m.mb_var_sum = 0;
enc->m.s.mc_mb_var_sum = coef_sum;
enc->m.s.mb_var_sum = 0;
}
pict->quality= ff_rate_estimate_qscale(&enc->m, 1);
@ -1757,7 +1759,7 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt,
{
SnowEncContext *const enc = avctx->priv_data;
SnowContext *const s = &enc->com;
MpegEncContext *const mpv = &enc->m;
MpegEncContext *const mpv = &enc->m.s;
RangeCoder * const c= &s->c;
AVCodecInternal *avci = avctx->internal;
AVFrame *pic;
@ -1793,10 +1795,10 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt,
mpv->picture_number = avctx->frame_num;
if(avctx->flags&AV_CODEC_FLAG_PASS2){
mpv->pict_type = pic->pict_type = mpv->rc_context.entry[avctx->frame_num].new_pict_type;
mpv->pict_type = pic->pict_type = enc->m.rc_context.entry[avctx->frame_num].new_pict_type;
s->keyframe = pic->pict_type == AV_PICTURE_TYPE_I;
if(!(avctx->flags&AV_CODEC_FLAG_QSCALE)) {
pic->quality = ff_rate_estimate_qscale(mpv, 0);
pic->quality = ff_rate_estimate_qscale(&enc->m, 0);
if (pic->quality < 0)
return -1;
}
@ -1877,7 +1879,7 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt,
mpv->qdsp = enc->qdsp; //move
mpv->hdsp = s->hdsp;
ff_me_init_pic(&enc->m);
ff_me_init_pic(mpv);
s->hdsp = mpv->hdsp;
}
@ -2043,17 +2045,19 @@ redo_frame:
s->current_picture->pict_type = pic->pict_type;
s->current_picture->quality = pic->quality;
mpv->frame_bits = 8 * (s->c.bytestream - s->c.bytestream_start);
mpv->p_tex_bits = mpv->frame_bits - mpv->misc_bits - mpv->mv_bits;
mpv->total_bits += 8*(s->c.bytestream - s->c.bytestream_start);
enc->m.frame_bits = 8 * (s->c.bytestream - s->c.bytestream_start);
mpv->p_tex_bits = enc->m.frame_bits - mpv->misc_bits - mpv->mv_bits;
enc->m.total_bits += 8*(s->c.bytestream - s->c.bytestream_start);
enc->cur_pic.display_picture_number =
enc->cur_pic.coded_picture_number = avctx->frame_num;
enc->cur_pic.f->quality = pic->quality;
if (enc->pass1_rc)
if (ff_rate_estimate_qscale(mpv, 0) < 0)
return -1;
if (enc->pass1_rc) {
ret = ff_rate_estimate_qscale(&enc->m, 0);
if (ret < 0)
return ret;
}
if(avctx->flags&AV_CODEC_FLAG_PASS1)
ff_write_pass1_stats(mpv);
ff_write_pass1_stats(&enc->m);
mpv->last_pict_type = mpv->pict_type;
emms_c();
@ -2088,10 +2092,10 @@ static av_cold int encode_end(AVCodecContext *avctx)
av_freep(&s->ref_scores[i]);
}
enc->m.me.temp = NULL;
av_freep(&enc->m.me.scratchpad);
av_freep(&enc->m.me.map);
av_freep(&enc->m.sc.obmc_scratchpad);
enc->m.s.me.temp = NULL;
av_freep(&enc->m.s.me.scratchpad);
av_freep(&enc->m.s.me.map);
av_freep(&enc->m.s.sc.obmc_scratchpad);
av_freep(&avctx->stats_out);