mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
mpegvideo_enc: K&R cosmetics (line 1000-2000).
Signed-off-by: Ronald S. Bultje <rsbultje@gmail.com>
This commit is contained in:
parent
7636c8c6a3
commit
9342ecf02e
@ -1140,7 +1140,8 @@ static int estimate_best_b_count(MpegEncContext *s)
|
||||
for (i = 0; i < s->max_b_frames + 2; i++) {
|
||||
int ysize = c->width * c->height;
|
||||
int csize = (c->width / 2) * (c->height / 2);
|
||||
Picture pre_input, *pre_input_ptr= i ? s->input_picture[i-1] : s->next_picture_ptr;
|
||||
Picture pre_input, *pre_input_ptr = i ? s->input_picture[i - 1] :
|
||||
s->next_picture_ptr;
|
||||
|
||||
avcodec_get_frame_defaults(&input[i]);
|
||||
input[i].data[0] = av_malloc(ysize + 2 * csize);
|
||||
@ -1159,9 +1160,15 @@ static int estimate_best_b_count(MpegEncContext *s)
|
||||
pre_input.f.data[2] += INPLACE_OFFSET;
|
||||
}
|
||||
|
||||
s->dsp.shrink[scale](input[i].data[0], input[i].linesize[0], pre_input.f.data[0], pre_input.f.linesize[0], c->width, c->height);
|
||||
s->dsp.shrink[scale](input[i].data[1], input[i].linesize[1], pre_input.f.data[1], pre_input.f.linesize[1], c->width >> 1, c->height >> 1);
|
||||
s->dsp.shrink[scale](input[i].data[2], input[i].linesize[2], pre_input.f.data[2], pre_input.f.linesize[2], c->width >> 1, c->height >> 1);
|
||||
s->dsp.shrink[scale](input[i].data[0], input[i].linesize[0],
|
||||
pre_input.f.data[0], pre_input.f.linesize[0],
|
||||
c->width, c->height);
|
||||
s->dsp.shrink[scale](input[i].data[1], input[i].linesize[1],
|
||||
pre_input.f.data[1], pre_input.f.linesize[1],
|
||||
c->width >> 1, c->height >> 1);
|
||||
s->dsp.shrink[scale](input[i].data[2], input[i].linesize[2],
|
||||
pre_input.f.data[2], pre_input.f.linesize[2],
|
||||
c->width >> 1, c->height >> 1);
|
||||
}
|
||||
}
|
||||
|
||||
@ -1175,15 +1182,18 @@ static int estimate_best_b_count(MpegEncContext *s)
|
||||
|
||||
input[0].pict_type = AV_PICTURE_TYPE_I;
|
||||
input[0].quality = 1 * FF_QP2LAMBDA;
|
||||
out_size = avcodec_encode_video(c, outbuf, outbuf_size, &input[0]);
|
||||
out_size = avcodec_encode_video(c, outbuf,
|
||||
outbuf_size, &input[0]);
|
||||
//rd += (out_size * lambda2) >> FF_LAMBDA_SHIFT;
|
||||
|
||||
for (i = 0; i < s->max_b_frames + 1; i++) {
|
||||
int is_p = i % (j + 1) == j || i == s->max_b_frames;
|
||||
|
||||
input[i+1].pict_type= is_p ? AV_PICTURE_TYPE_P : AV_PICTURE_TYPE_B;
|
||||
input[i + 1].pict_type = is_p ?
|
||||
AV_PICTURE_TYPE_P : AV_PICTURE_TYPE_B;
|
||||
input[i + 1].quality = is_p ? p_lambda : b_lambda;
|
||||
out_size = avcodec_encode_video(c, outbuf, outbuf_size, &input[i+1]);
|
||||
out_size = avcodec_encode_video(c, outbuf, outbuf_size,
|
||||
&input[i + 1]);
|
||||
rd += (out_size * lambda2) >> (FF_LAMBDA_SHIFT - 3);
|
||||
}
|
||||
|
||||
@ -1212,7 +1222,8 @@ static int estimate_best_b_count(MpegEncContext *s)
|
||||
return best_b_count;
|
||||
}
|
||||
|
||||
static int select_input_picture(MpegEncContext *s){
|
||||
static int select_input_picture(MpegEncContext *s)
|
||||
{
|
||||
int i;
|
||||
|
||||
for (i = 1; i < MAX_PICTURE_COUNT; i++)
|
||||
@ -1221,27 +1232,33 @@ static int select_input_picture(MpegEncContext *s){
|
||||
|
||||
/* set next picture type & ordering */
|
||||
if (s->reordered_input_picture[0] == NULL && s->input_picture[0]) {
|
||||
if(/*s->picture_in_gop_number >= s->gop_size ||*/ s->next_picture_ptr==NULL || s->intra_only){
|
||||
if (/*s->picture_in_gop_number >= s->gop_size ||*/
|
||||
s->next_picture_ptr == NULL || s->intra_only) {
|
||||
s->reordered_input_picture[0] = s->input_picture[0];
|
||||
s->reordered_input_picture[0]->f.pict_type = AV_PICTURE_TYPE_I;
|
||||
s->reordered_input_picture[0]->f.coded_picture_number = s->coded_picture_number++;
|
||||
s->reordered_input_picture[0]->f.coded_picture_number =
|
||||
s->coded_picture_number++;
|
||||
} else {
|
||||
int b_frames;
|
||||
|
||||
if (s->avctx->frame_skip_threshold || s->avctx->frame_skip_factor) {
|
||||
if(s->picture_in_gop_number < s->gop_size && skip_check(s, s->input_picture[0], s->next_picture_ptr)){
|
||||
if (s->picture_in_gop_number < s->gop_size &&
|
||||
skip_check(s, s->input_picture[0], s->next_picture_ptr)) {
|
||||
// FIXME check that te gop check above is +-1 correct
|
||||
//av_log(NULL, AV_LOG_DEBUG, "skip %p %"PRId64"\n", s->input_picture[0]->f.data[0], s->input_picture[0]->pts);
|
||||
//av_log(NULL, AV_LOG_DEBUG, "skip %p %"PRId64"\n",
|
||||
// s->input_picture[0]->f.data[0],
|
||||
// s->input_picture[0]->pts);
|
||||
|
||||
if (s->input_picture[0]->f.type == FF_BUFFER_TYPE_SHARED) {
|
||||
for (i = 0; i < 4; i++)
|
||||
s->input_picture[0]->f.data[i] = NULL;
|
||||
s->input_picture[0]->f.type = 0;
|
||||
} else {
|
||||
assert( s->input_picture[0]->f.type == FF_BUFFER_TYPE_USER
|
||||
|| s->input_picture[0]->f.type == FF_BUFFER_TYPE_INTERNAL);
|
||||
assert(s->input_picture[0]->f.type == FF_BUFFER_TYPE_USER ||
|
||||
s->input_picture[0]->f.type == FF_BUFFER_TYPE_INTERNAL);
|
||||
|
||||
s->avctx->release_buffer(s->avctx, (AVFrame*)s->input_picture[0]);
|
||||
s->avctx->release_buffer(s->avctx,
|
||||
(AVFrame *) s->input_picture[0]);
|
||||
}
|
||||
|
||||
emms_c();
|
||||
@ -1269,17 +1286,24 @@ static int select_input_picture(MpegEncContext *s){
|
||||
|
||||
if (s->avctx->b_frame_strategy == 0) {
|
||||
b_frames = s->max_b_frames;
|
||||
while(b_frames && !s->input_picture[b_frames]) b_frames--;
|
||||
while (b_frames && !s->input_picture[b_frames])
|
||||
b_frames--;
|
||||
} else if (s->avctx->b_frame_strategy == 1) {
|
||||
for (i = 1; i < s->max_b_frames + 1; i++) {
|
||||
if(s->input_picture[i] && s->input_picture[i]->b_frame_score==0){
|
||||
if (s->input_picture[i] &&
|
||||
s->input_picture[i]->b_frame_score == 0) {
|
||||
s->input_picture[i]->b_frame_score =
|
||||
get_intra_count(s, s->input_picture[i ]->f.data[0],
|
||||
s->input_picture[i-1]->f.data[0], s->linesize) + 1;
|
||||
get_intra_count(s,
|
||||
s->input_picture[i ]->f.data[0],
|
||||
s->input_picture[i - 1]->f.data[0],
|
||||
s->linesize) + 1;
|
||||
}
|
||||
}
|
||||
for (i = 0; i < s->max_b_frames + 1; i++) {
|
||||
if(s->input_picture[i]==NULL || s->input_picture[i]->b_frame_score - 1 > s->mb_num/s->avctx->b_sensitivity) break;
|
||||
if (s->input_picture[i] == NULL ||
|
||||
s->input_picture[i]->b_frame_score - 1 >
|
||||
s->mb_num / s->avctx->b_sensitivity)
|
||||
break;
|
||||
}
|
||||
|
||||
b_frames = FFMAX(0, i - 1);
|
||||
@ -1305,12 +1329,15 @@ static int select_input_picture(MpegEncContext *s){
|
||||
if (type && type != AV_PICTURE_TYPE_B)
|
||||
b_frames = i;
|
||||
}
|
||||
if (s->input_picture[b_frames]->f.pict_type == AV_PICTURE_TYPE_B && b_frames == s->max_b_frames){
|
||||
av_log(s->avctx, AV_LOG_ERROR, "warning, too many b frames in a row\n");
|
||||
if (s->input_picture[b_frames]->f.pict_type == AV_PICTURE_TYPE_B &&
|
||||
b_frames == s->max_b_frames) {
|
||||
av_log(s->avctx, AV_LOG_ERROR,
|
||||
"warning, too many b frames in a row\n");
|
||||
}
|
||||
|
||||
if (s->picture_in_gop_number + b_frames >= s->gop_size) {
|
||||
if((s->flags2 & CODEC_FLAG2_STRICT_GOP) && s->gop_size > s->picture_in_gop_number){
|
||||
if ((s->flags2 & CODEC_FLAG2_STRICT_GOP) &&
|
||||
s->gop_size > s->picture_in_gop_number) {
|
||||
b_frames = s->gop_size - s->picture_in_gop_number - 1;
|
||||
} else {
|
||||
if (s->flags & CODEC_FLAG_CLOSED_GOP)
|
||||
@ -1319,30 +1346,36 @@ static int select_input_picture(MpegEncContext *s){
|
||||
}
|
||||
}
|
||||
|
||||
if( (s->flags & CODEC_FLAG_CLOSED_GOP)
|
||||
&& b_frames
|
||||
&& s->input_picture[b_frames]->f.pict_type== AV_PICTURE_TYPE_I)
|
||||
if ((s->flags & CODEC_FLAG_CLOSED_GOP) && b_frames &&
|
||||
s->input_picture[b_frames]->f.pict_type == AV_PICTURE_TYPE_I)
|
||||
b_frames--;
|
||||
|
||||
s->reordered_input_picture[0] = s->input_picture[b_frames];
|
||||
if (s->reordered_input_picture[0]->f.pict_type != AV_PICTURE_TYPE_I)
|
||||
s->reordered_input_picture[0]->f.pict_type = AV_PICTURE_TYPE_P;
|
||||
s->reordered_input_picture[0]->f.coded_picture_number = s->coded_picture_number++;
|
||||
s->reordered_input_picture[0]->f.coded_picture_number =
|
||||
s->coded_picture_number++;
|
||||
for (i = 0; i < b_frames; i++) {
|
||||
s->reordered_input_picture[i + 1] = s->input_picture[i];
|
||||
s->reordered_input_picture[i + 1]->f.pict_type = AV_PICTURE_TYPE_B;
|
||||
s->reordered_input_picture[i + 1]->f.coded_picture_number = s->coded_picture_number++;
|
||||
s->reordered_input_picture[i + 1]->f.pict_type =
|
||||
AV_PICTURE_TYPE_B;
|
||||
s->reordered_input_picture[i + 1]->f.coded_picture_number =
|
||||
s->coded_picture_number++;
|
||||
}
|
||||
}
|
||||
}
|
||||
no_output_pic:
|
||||
if (s->reordered_input_picture[0]) {
|
||||
s->reordered_input_picture[0]->f.reference = s->reordered_input_picture[0]->f.pict_type!=AV_PICTURE_TYPE_B ? 3 : 0;
|
||||
s->reordered_input_picture[0]->f.reference =
|
||||
s->reordered_input_picture[0]->f.pict_type !=
|
||||
AV_PICTURE_TYPE_B ? 3 : 0;
|
||||
|
||||
ff_copy_picture(&s->new_picture, s->reordered_input_picture[0]);
|
||||
|
||||
if (s->reordered_input_picture[0]->f.type == FF_BUFFER_TYPE_SHARED || s->avctx->rc_buffer_size) {
|
||||
// input is a shared pix, so we can't modifiy it -> alloc a new one & ensure that the shared one is reuseable
|
||||
if (s->reordered_input_picture[0]->f.type == FF_BUFFER_TYPE_SHARED ||
|
||||
s->avctx->rc_buffer_size) {
|
||||
// input is a shared pix, so we can't modifiy it -> alloc a new
|
||||
// one & ensure that the shared one is reuseable
|
||||
|
||||
Picture *pic;
|
||||
int i = ff_find_unused_picture(s, 0);
|
||||
@ -1357,19 +1390,23 @@ no_output_pic:
|
||||
|
||||
/* mark us unused / free shared pic */
|
||||
if (s->reordered_input_picture[0]->f.type == FF_BUFFER_TYPE_INTERNAL)
|
||||
s->avctx->release_buffer(s->avctx, (AVFrame*)s->reordered_input_picture[0]);
|
||||
s->avctx->release_buffer(s->avctx,
|
||||
(AVFrame *) s->reordered_input_picture[0]);
|
||||
for (i = 0; i < 4; i++)
|
||||
s->reordered_input_picture[0]->f.data[i] = NULL;
|
||||
s->reordered_input_picture[0]->f.type = 0;
|
||||
|
||||
copy_picture_attributes(s, (AVFrame*)pic, (AVFrame*)s->reordered_input_picture[0]);
|
||||
copy_picture_attributes(s, (AVFrame *) pic,
|
||||
(AVFrame *) s->reordered_input_picture[0]);
|
||||
|
||||
s->current_picture_ptr = pic;
|
||||
} else {
|
||||
// input is not a shared pix -> reuse buffer for current_pix
|
||||
|
||||
assert( s->reordered_input_picture[0]->f.type == FF_BUFFER_TYPE_USER
|
||||
|| s->reordered_input_picture[0]->f.type == FF_BUFFER_TYPE_INTERNAL);
|
||||
assert(s->reordered_input_picture[0]->f.type ==
|
||||
FF_BUFFER_TYPE_USER ||
|
||||
s->reordered_input_picture[0]->f.type ==
|
||||
FF_BUFFER_TYPE_INTERNAL);
|
||||
|
||||
s->current_picture_ptr = s->reordered_input_picture[0];
|
||||
for (i = 0; i < 4; i++) {
|
||||
@ -1416,7 +1453,8 @@ int MPV_encode_picture(AVCodecContext *avctx,
|
||||
if (s->new_picture.f.data[0]) {
|
||||
s->pict_type = s->new_picture.f.pict_type;
|
||||
//emms_c();
|
||||
//printf("qs:%f %f %d\n", s->new_picture.quality, s->current_picture.quality, s->qscale);
|
||||
//printf("qs:%f %f %d\n", s->new_picture.quality,
|
||||
// s->current_picture.quality, s->qscale);
|
||||
MPV_frame_start(s, avctx);
|
||||
vbv_retry:
|
||||
if (encode_picture(s, s->picture_number) < 0)
|
||||
@ -1428,7 +1466,8 @@ vbv_retry:
|
||||
avctx->i_tex_bits = s->i_tex_bits;
|
||||
avctx->p_tex_bits = s->p_tex_bits;
|
||||
avctx->i_count = s->i_count;
|
||||
avctx->p_count = s->mb_num - s->i_count - s->skip_count; //FIXME f/b_count in avctx
|
||||
// FIXME f/b_count in avctx
|
||||
avctx->p_count = s->mb_num - s->i_count - s->skip_count;
|
||||
avctx->skip_count = s->skip_count;
|
||||
|
||||
MPV_frame_end(s);
|
||||
@ -1440,16 +1479,24 @@ vbv_retry:
|
||||
RateControlContext *rcc = &s->rc_context;
|
||||
int max_size = rcc->buffer_index * avctx->rc_max_available_vbv_use;
|
||||
|
||||
if(put_bits_count(&s->pb) > max_size && s->lambda < s->avctx->lmax){
|
||||
s->next_lambda= FFMAX(s->lambda+1, s->lambda*(s->qscale+1) / s->qscale);
|
||||
if (put_bits_count(&s->pb) > max_size &&
|
||||
s->lambda < s->avctx->lmax) {
|
||||
s->next_lambda = FFMAX(s->lambda + 1, s->lambda *
|
||||
(s->qscale + 1) / s->qscale);
|
||||
if (s->adaptive_quant) {
|
||||
int i;
|
||||
for (i = 0; i < s->mb_height * s->mb_stride; i++)
|
||||
s->lambda_table[i]= FFMAX(s->lambda_table[i]+1, s->lambda_table[i]*(s->qscale+1) / s->qscale);
|
||||
s->lambda_table[i] =
|
||||
FFMAX(s->lambda_table[i] + 1,
|
||||
s->lambda_table[i] * (s->qscale + 1) /
|
||||
s->qscale);
|
||||
}
|
||||
s->mb_skipped = 0; // done in MPV_frame_start()
|
||||
if(s->pict_type==AV_PICTURE_TYPE_P){ //done in encode_picture() so we must undo it
|
||||
if(s->flipflop_rounding || s->codec_id == CODEC_ID_H263P || s->codec_id == CODEC_ID_MPEG4)
|
||||
// done in encode_picture() so we must undo it
|
||||
if (s->pict_type == AV_PICTURE_TYPE_P) {
|
||||
if (s->flipflop_rounding ||
|
||||
s->codec_id == CODEC_ID_H263P ||
|
||||
s->codec_id == CODEC_ID_MPEG4)
|
||||
s->no_rounding ^= 1;
|
||||
}
|
||||
if (s->pict_type != AV_PICTURE_TYPE_B) {
|
||||
@ -1476,13 +1523,16 @@ vbv_retry:
|
||||
}
|
||||
|
||||
if (s->flags & CODEC_FLAG_PASS1)
|
||||
assert(avctx->header_bits + avctx->mv_bits + avctx->misc_bits + avctx->i_tex_bits + avctx->p_tex_bits == put_bits_count(&s->pb));
|
||||
assert(avctx->header_bits + avctx->mv_bits + avctx->misc_bits +
|
||||
avctx->i_tex_bits + avctx->p_tex_bits ==
|
||||
put_bits_count(&s->pb));
|
||||
flush_put_bits(&s->pb);
|
||||
s->frame_bits = put_bits_count(&s->pb);
|
||||
|
||||
stuffing_count = ff_vbv_update(s, s->frame_bits);
|
||||
if (stuffing_count) {
|
||||
if(s->pb.buf_end - s->pb.buf - (put_bits_count(&s->pb)>>3) < stuffing_count + 50){
|
||||
if (s->pb.buf_end - s->pb.buf - (put_bits_count(&s->pb) >> 3) <
|
||||
stuffing_count + 50) {
|
||||
av_log(s->avctx, AV_LOG_ERROR, "stuffing too large\n");
|
||||
return -1;
|
||||
}
|
||||
@ -1510,20 +1560,27 @@ vbv_retry:
|
||||
}
|
||||
|
||||
/* update mpeg1/2 vbv_delay for CBR */
|
||||
if(s->avctx->rc_max_rate && s->avctx->rc_min_rate == s->avctx->rc_max_rate && s->out_format == FMT_MPEG1
|
||||
&& 90000LL * (avctx->rc_buffer_size-1) <= s->avctx->rc_max_rate*0xFFFFLL){
|
||||
if (s->avctx->rc_max_rate &&
|
||||
s->avctx->rc_min_rate == s->avctx->rc_max_rate &&
|
||||
s->out_format == FMT_MPEG1 &&
|
||||
90000LL * (avctx->rc_buffer_size - 1) <=
|
||||
s->avctx->rc_max_rate * 0xFFFFLL) {
|
||||
int vbv_delay, min_delay;
|
||||
double inbits = s->avctx->rc_max_rate*av_q2d(s->avctx->time_base);
|
||||
int minbits= s->frame_bits - 8*(s->vbv_delay_ptr - s->pb.buf - 1);
|
||||
double inbits = s->avctx->rc_max_rate *
|
||||
av_q2d(s->avctx->time_base);
|
||||
int minbits = s->frame_bits - 8 *
|
||||
(s->vbv_delay_ptr - s->pb.buf - 1);
|
||||
double bits = s->rc_context.buffer_index + minbits - inbits;
|
||||
|
||||
if (bits < 0)
|
||||
av_log(s->avctx, AV_LOG_ERROR, "Internal error, negative bits\n");
|
||||
av_log(s->avctx, AV_LOG_ERROR,
|
||||
"Internal error, negative bits\n");
|
||||
|
||||
assert(s->repeat_first_field == 0);
|
||||
|
||||
vbv_delay = bits * 90000 / s->avctx->rc_max_rate;
|
||||
min_delay= (minbits * 90000LL + s->avctx->rc_max_rate - 1)/ s->avctx->rc_max_rate;
|
||||
min_delay = (minbits * 90000LL + s->avctx->rc_max_rate - 1) /
|
||||
s->avctx->rc_max_rate;
|
||||
|
||||
vbv_delay = FFMAX(vbv_delay, min_delay);
|
||||
|
||||
@ -1547,17 +1604,19 @@ vbv_retry:
|
||||
return s->frame_bits / 8;
|
||||
}
|
||||
|
||||
static inline void dct_single_coeff_elimination(MpegEncContext *s, int n, int threshold)
|
||||
static inline void dct_single_coeff_elimination(MpegEncContext *s,
|
||||
int n, int threshold)
|
||||
{
|
||||
static const char tab[64]=
|
||||
{3,2,2,1,1,1,1,1,
|
||||
static const char tab[64] = {
|
||||
3, 2, 2, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0,0,0,0,0,0,0,0};
|
||||
0, 0, 0, 0, 0, 0, 0, 0
|
||||
};
|
||||
int score = 0;
|
||||
int run = 0;
|
||||
int i;
|
||||
@ -1572,13 +1631,15 @@ static inline void dct_single_coeff_elimination(MpegEncContext *s, int n, int th
|
||||
skip_dc = 1;
|
||||
|
||||
/* Are all we could set to zero already zero? */
|
||||
if(last_index<=skip_dc - 1) return;
|
||||
if (last_index <= skip_dc - 1)
|
||||
return;
|
||||
|
||||
for (i = 0; i <= last_index; i++) {
|
||||
const int j = s->intra_scantable.permutated[i];
|
||||
const int level = FFABS(block[j]);
|
||||
if (level == 1) {
|
||||
if(skip_dc && i==0) continue;
|
||||
if (skip_dc && i == 0)
|
||||
continue;
|
||||
score += tab[run];
|
||||
run = 0;
|
||||
} else if (level > 1) {
|
||||
@ -1587,16 +1648,20 @@ static inline void dct_single_coeff_elimination(MpegEncContext *s, int n, int th
|
||||
run++;
|
||||
}
|
||||
}
|
||||
if(score >= threshold) return;
|
||||
if (score >= threshold)
|
||||
return;
|
||||
for (i = skip_dc; i <= last_index; i++) {
|
||||
const int j = s->intra_scantable.permutated[i];
|
||||
block[j] = 0;
|
||||
}
|
||||
if(block[0]) s->block_last_index[n]= 0;
|
||||
else s->block_last_index[n]= -1;
|
||||
if (block[0])
|
||||
s->block_last_index[n] = 0;
|
||||
else
|
||||
s->block_last_index[n] = -1;
|
||||
}
|
||||
|
||||
static inline void clip_coeffs(MpegEncContext *s, DCTELEM *block, int last_index)
|
||||
static inline void clip_coeffs(MpegEncContext *s, DCTELEM *block,
|
||||
int last_index)
|
||||
{
|
||||
int i;
|
||||
const int maxlevel = s->max_qcoeff;
|
||||
@ -1624,10 +1689,13 @@ static inline void clip_coeffs(MpegEncContext *s, DCTELEM *block, int last_index
|
||||
}
|
||||
|
||||
if (overflow && s->avctx->mb_decision == FF_MB_DECISION_SIMPLE)
|
||||
av_log(s->avctx, AV_LOG_INFO, "warning, clipping %d dct coefficients to %d..%d\n", overflow, minlevel, maxlevel);
|
||||
av_log(s->avctx, AV_LOG_INFO,
|
||||
"warning, clipping %d dct coefficients to %d..%d\n",
|
||||
overflow, minlevel, maxlevel);
|
||||
}
|
||||
|
||||
static void get_visual_weight(int16_t *weight, uint8_t *ptr, int stride){
|
||||
static void get_visual_weight(int16_t *weight, uint8_t *ptr, int stride)
|
||||
{
|
||||
int x, y;
|
||||
// FIXME optimize
|
||||
for (y = 0; y < 8; y++) {
|
||||
@ -1650,7 +1718,10 @@ static void get_visual_weight(int16_t *weight, uint8_t *ptr, int stride){
|
||||
}
|
||||
}
|
||||
|
||||
static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x, int motion_y, int mb_block_height, int mb_block_count)
|
||||
static av_always_inline void encode_mb_internal(MpegEncContext *s,
|
||||
int motion_x, int motion_y,
|
||||
int mb_block_height,
|
||||
int mb_block_count)
|
||||
{
|
||||
int16_t weight[8][64];
|
||||
DCTELEM orig[8][64];
|
||||
@ -1662,7 +1733,8 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
uint8_t *ptr_y, *ptr_cb, *ptr_cr;
|
||||
int wrap_y, wrap_c;
|
||||
|
||||
for(i=0; i<mb_block_count; i++) skip_dct[i]=s->skipdct;
|
||||
for (i = 0; i < mb_block_count; i++)
|
||||
skip_dct[i] = s->skipdct;
|
||||
|
||||
if (s->adaptive_quant) {
|
||||
const int last_qp = s->qscale;
|
||||
@ -1696,17 +1768,25 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
|
||||
wrap_y = s->linesize;
|
||||
wrap_c = s->uvlinesize;
|
||||
ptr_y = s->new_picture.f.data[0] + (mb_y * 16 * wrap_y) + mb_x * 16;
|
||||
ptr_cb = s->new_picture.f.data[1] + (mb_y * mb_block_height * wrap_c) + mb_x * 8;
|
||||
ptr_cr = s->new_picture.f.data[2] + (mb_y * mb_block_height * wrap_c) + mb_x * 8;
|
||||
ptr_y = s->new_picture.f.data[0] +
|
||||
(mb_y * 16 * wrap_y) + mb_x * 16;
|
||||
ptr_cb = s->new_picture.f.data[1] +
|
||||
(mb_y * mb_block_height * wrap_c) + mb_x * 8;
|
||||
ptr_cr = s->new_picture.f.data[2] +
|
||||
(mb_y * mb_block_height * wrap_c) + mb_x * 8;
|
||||
|
||||
if (mb_x * 16 + 16 > s->width || mb_y * 16 + 16 > s->height) {
|
||||
uint8_t *ebuf = s->edge_emu_buffer + 32;
|
||||
s->dsp.emulated_edge_mc(ebuf , ptr_y , wrap_y,16,16,mb_x*16,mb_y*16, s->width , s->height);
|
||||
s->dsp.emulated_edge_mc(ebuf, ptr_y, wrap_y, 16, 16, mb_x * 16,
|
||||
mb_y * 16, s->width, s->height);
|
||||
ptr_y = ebuf;
|
||||
s->dsp.emulated_edge_mc(ebuf+18*wrap_y , ptr_cb, wrap_c, 8, mb_block_height, mb_x*8, mb_y*8, s->width>>1, s->height>>1);
|
||||
s->dsp.emulated_edge_mc(ebuf + 18 * wrap_y, ptr_cb, wrap_c, 8,
|
||||
mb_block_height, mb_x * 8, mb_y * 8,
|
||||
s->width >> 1, s->height >> 1);
|
||||
ptr_cb = ebuf + 18 * wrap_y;
|
||||
s->dsp.emulated_edge_mc(ebuf+18*wrap_y+8, ptr_cr, wrap_c, 8, mb_block_height, mb_x*8, mb_y*8, s->width>>1, s->height>>1);
|
||||
s->dsp.emulated_edge_mc(ebuf + 18 * wrap_y + 8, ptr_cr, wrap_c, 8,
|
||||
mb_block_height, mb_x * 8, mb_y * 8,
|
||||
s->width >> 1, s->height >> 1);
|
||||
ptr_cr = ebuf + 18 * wrap_y + 8;
|
||||
}
|
||||
|
||||
@ -1715,12 +1795,16 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
int progressive_score, interlaced_score;
|
||||
|
||||
s->interlaced_dct = 0;
|
||||
progressive_score= s->dsp.ildct_cmp[4](s, ptr_y , NULL, wrap_y, 8)
|
||||
+s->dsp.ildct_cmp[4](s, ptr_y + wrap_y*8, NULL, wrap_y, 8) - 400;
|
||||
progressive_score = s->dsp.ildct_cmp[4](s, ptr_y,
|
||||
NULL, wrap_y, 8) +
|
||||
s->dsp.ildct_cmp[4](s, ptr_y + wrap_y * 8,
|
||||
NULL, wrap_y, 8) - 400;
|
||||
|
||||
if (progressive_score > 0) {
|
||||
interlaced_score = s->dsp.ildct_cmp[4](s, ptr_y , NULL, wrap_y*2, 8)
|
||||
+s->dsp.ildct_cmp[4](s, ptr_y + wrap_y , NULL, wrap_y*2, 8);
|
||||
interlaced_score = s->dsp.ildct_cmp[4](s, ptr_y,
|
||||
NULL, wrap_y * 2, 8) +
|
||||
s->dsp.ildct_cmp[4](s, ptr_y + wrap_y,
|
||||
NULL, wrap_y * 2, 8);
|
||||
if (progressive_score > interlaced_score) {
|
||||
s->interlaced_dct = 1;
|
||||
|
||||
@ -1744,8 +1828,10 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
s->dsp.get_pixels(s->block[4], ptr_cb, wrap_c);
|
||||
s->dsp.get_pixels(s->block[5], ptr_cr, wrap_c);
|
||||
if (!s->chroma_y_shift) { /* 422 */
|
||||
s->dsp.get_pixels(s->block[6], ptr_cb + (dct_offset>>1), wrap_c);
|
||||
s->dsp.get_pixels(s->block[7], ptr_cr + (dct_offset>>1), wrap_c);
|
||||
s->dsp.get_pixels(s->block[6],
|
||||
ptr_cb + (dct_offset >> 1), wrap_c);
|
||||
s->dsp.get_pixels(s->block[7],
|
||||
ptr_cr + (dct_offset >> 1), wrap_c);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
@ -1766,26 +1852,37 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
}
|
||||
|
||||
if (s->mv_dir & MV_DIR_FORWARD) {
|
||||
MPV_motion(s, dest_y, dest_cb, dest_cr, 0, s->last_picture.f.data, op_pix, op_qpix);
|
||||
MPV_motion(s, dest_y, dest_cb, dest_cr, 0, s->last_picture.f.data,
|
||||
op_pix, op_qpix);
|
||||
op_pix = s->dsp.avg_pixels_tab;
|
||||
op_qpix = s->dsp.avg_qpel_pixels_tab;
|
||||
}
|
||||
if (s->mv_dir & MV_DIR_BACKWARD) {
|
||||
MPV_motion(s, dest_y, dest_cb, dest_cr, 1, s->next_picture.f.data, op_pix, op_qpix);
|
||||
MPV_motion(s, dest_y, dest_cb, dest_cr, 1, s->next_picture.f.data,
|
||||
op_pix, op_qpix);
|
||||
}
|
||||
|
||||
if (s->flags & CODEC_FLAG_INTERLACED_DCT) {
|
||||
int progressive_score, interlaced_score;
|
||||
|
||||
s->interlaced_dct = 0;
|
||||
progressive_score= s->dsp.ildct_cmp[0](s, dest_y , ptr_y , wrap_y, 8)
|
||||
+s->dsp.ildct_cmp[0](s, dest_y + wrap_y*8, ptr_y + wrap_y*8, wrap_y, 8) - 400;
|
||||
progressive_score = s->dsp.ildct_cmp[0](s, dest_y,
|
||||
ptr_y, wrap_y,
|
||||
8) +
|
||||
s->dsp.ildct_cmp[0](s, dest_y + wrap_y * 8,
|
||||
ptr_y + wrap_y * 8, wrap_y,
|
||||
8) - 400;
|
||||
|
||||
if(s->avctx->ildct_cmp == FF_CMP_VSSE) progressive_score -= 400;
|
||||
if (s->avctx->ildct_cmp == FF_CMP_VSSE)
|
||||
progressive_score -= 400;
|
||||
|
||||
if (progressive_score > 0) {
|
||||
interlaced_score = s->dsp.ildct_cmp[0](s, dest_y , ptr_y , wrap_y*2, 8)
|
||||
+s->dsp.ildct_cmp[0](s, dest_y + wrap_y , ptr_y + wrap_y , wrap_y*2, 8);
|
||||
interlaced_score = s->dsp.ildct_cmp[0](s, dest_y,
|
||||
ptr_y,
|
||||
wrap_y * 2, 8) +
|
||||
s->dsp.ildct_cmp[0](s, dest_y + wrap_y,
|
||||
ptr_y + wrap_y,
|
||||
wrap_y * 2, 8);
|
||||
|
||||
if (progressive_score > interlaced_score) {
|
||||
s->interlaced_dct = 1;
|
||||
@ -1800,8 +1897,10 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
|
||||
s->dsp.diff_pixels(s->block[0], ptr_y, dest_y, wrap_y);
|
||||
s->dsp.diff_pixels(s->block[1], ptr_y + 8, dest_y + 8, wrap_y);
|
||||
s->dsp.diff_pixels(s->block[2], ptr_y + dct_offset , dest_y + dct_offset , wrap_y);
|
||||
s->dsp.diff_pixels(s->block[3], ptr_y + dct_offset + 8, dest_y + dct_offset + 8, wrap_y);
|
||||
s->dsp.diff_pixels(s->block[2], ptr_y + dct_offset,
|
||||
dest_y + dct_offset, wrap_y);
|
||||
s->dsp.diff_pixels(s->block[3], ptr_y + dct_offset + 8,
|
||||
dest_y + dct_offset + 8, wrap_y);
|
||||
|
||||
if (s->flags & CODEC_FLAG_GRAY) {
|
||||
skip_dct[4] = 1;
|
||||
@ -1810,36 +1909,68 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
s->dsp.diff_pixels(s->block[4], ptr_cb, dest_cb, wrap_c);
|
||||
s->dsp.diff_pixels(s->block[5], ptr_cr, dest_cr, wrap_c);
|
||||
if (!s->chroma_y_shift) { /* 422 */
|
||||
s->dsp.diff_pixels(s->block[6], ptr_cb + (dct_offset>>1), dest_cb + (dct_offset>>1), wrap_c);
|
||||
s->dsp.diff_pixels(s->block[7], ptr_cr + (dct_offset>>1), dest_cr + (dct_offset>>1), wrap_c);
|
||||
s->dsp.diff_pixels(s->block[6], ptr_cb + (dct_offset >> 1),
|
||||
dest_cb + (dct_offset >> 1), wrap_c);
|
||||
s->dsp.diff_pixels(s->block[7], ptr_cr + (dct_offset >> 1),
|
||||
dest_cr + (dct_offset >> 1), wrap_c);
|
||||
}
|
||||
}
|
||||
/* pre quantization */
|
||||
if(s->current_picture.mc_mb_var[s->mb_stride*mb_y+ mb_x]<2*s->qscale*s->qscale){
|
||||
if (s->current_picture.mc_mb_var[s->mb_stride * mb_y + mb_x] <
|
||||
2 * s->qscale * s->qscale) {
|
||||
// FIXME optimize
|
||||
if(s->dsp.sad[1](NULL, ptr_y , dest_y , wrap_y, 8) < 20*s->qscale) skip_dct[0]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_y + 8, dest_y + 8, wrap_y, 8) < 20*s->qscale) skip_dct[1]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_y +dct_offset , dest_y +dct_offset , wrap_y, 8) < 20*s->qscale) skip_dct[2]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_y +dct_offset+ 8, dest_y +dct_offset+ 8, wrap_y, 8) < 20*s->qscale) skip_dct[3]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_cb , dest_cb , wrap_c, 8) < 20*s->qscale) skip_dct[4]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_cr , dest_cr , wrap_c, 8) < 20*s->qscale) skip_dct[5]= 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_y , dest_y,
|
||||
wrap_y, 8) < 20 * s->qscale)
|
||||
skip_dct[0] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_y + 8,
|
||||
dest_y + 8, wrap_y, 8) < 20 * s->qscale)
|
||||
skip_dct[1] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_y + dct_offset,
|
||||
dest_y + dct_offset, wrap_y, 8) < 20 * s->qscale)
|
||||
skip_dct[2] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_y + dct_offset + 8,
|
||||
dest_y + dct_offset + 8,
|
||||
wrap_y, 8) < 20 * s->qscale)
|
||||
skip_dct[3] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_cb, dest_cb,
|
||||
wrap_c, 8) < 20 * s->qscale)
|
||||
skip_dct[4] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_cr, dest_cr,
|
||||
wrap_c, 8) < 20 * s->qscale)
|
||||
skip_dct[5] = 1;
|
||||
if (!s->chroma_y_shift) { /* 422 */
|
||||
if(s->dsp.sad[1](NULL, ptr_cb +(dct_offset>>1), dest_cb +(dct_offset>>1), wrap_c, 8) < 20*s->qscale) skip_dct[6]= 1;
|
||||
if(s->dsp.sad[1](NULL, ptr_cr +(dct_offset>>1), dest_cr +(dct_offset>>1), wrap_c, 8) < 20*s->qscale) skip_dct[7]= 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_cb + (dct_offset >> 1),
|
||||
dest_cb + (dct_offset >> 1),
|
||||
wrap_c, 8) < 20 * s->qscale)
|
||||
skip_dct[6] = 1;
|
||||
if (s->dsp.sad[1](NULL, ptr_cr + (dct_offset >> 1),
|
||||
dest_cr + (dct_offset >> 1),
|
||||
wrap_c, 8) < 20 * s->qscale)
|
||||
skip_dct[7] = 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (s->avctx->quantizer_noise_shaping) {
|
||||
if(!skip_dct[0]) get_visual_weight(weight[0], ptr_y , wrap_y);
|
||||
if(!skip_dct[1]) get_visual_weight(weight[1], ptr_y + 8, wrap_y);
|
||||
if(!skip_dct[2]) get_visual_weight(weight[2], ptr_y + dct_offset , wrap_y);
|
||||
if(!skip_dct[3]) get_visual_weight(weight[3], ptr_y + dct_offset + 8, wrap_y);
|
||||
if(!skip_dct[4]) get_visual_weight(weight[4], ptr_cb , wrap_c);
|
||||
if(!skip_dct[5]) get_visual_weight(weight[5], ptr_cr , wrap_c);
|
||||
if (!skip_dct[0])
|
||||
get_visual_weight(weight[0], ptr_y , wrap_y);
|
||||
if (!skip_dct[1])
|
||||
get_visual_weight(weight[1], ptr_y + 8, wrap_y);
|
||||
if (!skip_dct[2])
|
||||
get_visual_weight(weight[2], ptr_y + dct_offset , wrap_y);
|
||||
if (!skip_dct[3])
|
||||
get_visual_weight(weight[3], ptr_y + dct_offset + 8, wrap_y);
|
||||
if (!skip_dct[4])
|
||||
get_visual_weight(weight[4], ptr_cb , wrap_c);
|
||||
if (!skip_dct[5])
|
||||
get_visual_weight(weight[5], ptr_cr , wrap_c);
|
||||
if (!s->chroma_y_shift) { /* 422 */
|
||||
if(!skip_dct[6]) get_visual_weight(weight[6], ptr_cb + (dct_offset>>1), wrap_c);
|
||||
if(!skip_dct[7]) get_visual_weight(weight[7], ptr_cr + (dct_offset>>1), wrap_c);
|
||||
if (!skip_dct[6])
|
||||
get_visual_weight(weight[6], ptr_cb + (dct_offset >> 1),
|
||||
wrap_c);
|
||||
if (!skip_dct[7])
|
||||
get_visual_weight(weight[7], ptr_cr + (dct_offset >> 1),
|
||||
wrap_c);
|
||||
}
|
||||
memcpy(orig[0], s->block[0], sizeof(DCTELEM) * 64 * mb_block_count);
|
||||
}
|
||||
@ -1851,17 +1982,22 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
if (!skip_dct[i]) {
|
||||
int overflow;
|
||||
s->block_last_index[i] = s->dct_quantize(s, s->block[i], i, s->qscale, &overflow);
|
||||
// FIXME we could decide to change to quantizer instead of clipping
|
||||
// JS: I don't think that would be a good idea it could lower quality instead
|
||||
// of improve it. Just INTRADC clipping deserves changes in quantizer
|
||||
if (overflow) clip_coeffs(s, s->block[i], s->block_last_index[i]);
|
||||
// FIXME we could decide to change to quantizer instead of
|
||||
// clipping
|
||||
// JS: I don't think that would be a good idea it could lower
|
||||
// quality instead of improve it. Just INTRADC clipping
|
||||
// deserves changes in quantizer
|
||||
if (overflow)
|
||||
clip_coeffs(s, s->block[i], s->block_last_index[i]);
|
||||
} else
|
||||
s->block_last_index[i] = -1;
|
||||
}
|
||||
if (s->avctx->quantizer_noise_shaping) {
|
||||
for (i = 0; i < mb_block_count; i++) {
|
||||
if (!skip_dct[i]) {
|
||||
s->block_last_index[i] = dct_quantize_refine(s, s->block[i], weight[i], orig[i], i, s->qscale);
|
||||
s->block_last_index[i] =
|
||||
dct_quantize_refine(s, s->block[i], weight[i],
|
||||
orig[i], i, s->qscale);
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -1894,7 +2030,8 @@ static av_always_inline void encode_mb_internal(MpegEncContext *s, int motion_x,
|
||||
int j;
|
||||
if (s->block_last_index[i] > 0) {
|
||||
for (j = 63; j > 0; j--) {
|
||||
if(s->block[i][ s->intra_scantable.permutated[j] ]) break;
|
||||
if (s->block[i][s->intra_scantable.permutated[j]])
|
||||
break;
|
||||
}
|
||||
s->block_last_index[i] = j;
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user