mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
Remove unused variables
This commit is contained in:
parent
808d8ff6bb
commit
e65ab9d94f
8
ffmpeg.c
8
ffmpeg.c
@ -1117,7 +1117,7 @@ static void do_video_out(AVFormatContext *s,
|
|||||||
int *frame_size)
|
int *frame_size)
|
||||||
{
|
{
|
||||||
int nb_frames, i, ret, resample_changed;
|
int nb_frames, i, ret, resample_changed;
|
||||||
AVFrame *final_picture, *formatted_picture, *resampling_dst;
|
AVFrame *final_picture, *formatted_picture;
|
||||||
AVCodecContext *enc, *dec;
|
AVCodecContext *enc, *dec;
|
||||||
double sync_ipts;
|
double sync_ipts;
|
||||||
|
|
||||||
@ -1162,7 +1162,6 @@ static void do_video_out(AVFormatContext *s,
|
|||||||
|
|
||||||
formatted_picture = in_picture;
|
formatted_picture = in_picture;
|
||||||
final_picture = formatted_picture;
|
final_picture = formatted_picture;
|
||||||
resampling_dst = &ost->pict_tmp;
|
|
||||||
|
|
||||||
resample_changed = ost->resample_width != dec->width ||
|
resample_changed = ost->resample_width != dec->width ||
|
||||||
ost->resample_height != dec->height ||
|
ost->resample_height != dec->height ||
|
||||||
@ -1198,7 +1197,7 @@ static void do_video_out(AVFormatContext *s,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
sws_scale(ost->img_resample_ctx, formatted_picture->data, formatted_picture->linesize,
|
sws_scale(ost->img_resample_ctx, formatted_picture->data, formatted_picture->linesize,
|
||||||
0, ost->resample_height, resampling_dst->data, resampling_dst->linesize);
|
0, ost->resample_height, final_picture->data, final_picture->linesize);
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
@ -3617,7 +3616,6 @@ static void new_audio_stream(AVFormatContext *oc, int file_idx)
|
|||||||
static void new_data_stream(AVFormatContext *oc, int file_idx)
|
static void new_data_stream(AVFormatContext *oc, int file_idx)
|
||||||
{
|
{
|
||||||
AVStream *st;
|
AVStream *st;
|
||||||
AVOutputStream *ost;
|
|
||||||
AVCodec *codec=NULL;
|
AVCodec *codec=NULL;
|
||||||
AVCodecContext *data_enc;
|
AVCodecContext *data_enc;
|
||||||
|
|
||||||
@ -3626,7 +3624,7 @@ static void new_data_stream(AVFormatContext *oc, int file_idx)
|
|||||||
fprintf(stderr, "Could not alloc stream\n");
|
fprintf(stderr, "Could not alloc stream\n");
|
||||||
ffmpeg_exit(1);
|
ffmpeg_exit(1);
|
||||||
}
|
}
|
||||||
ost = new_output_stream(oc, file_idx);
|
new_output_stream(oc, file_idx);
|
||||||
data_enc = st->codec;
|
data_enc = st->codec;
|
||||||
output_codecs = grow_array(output_codecs, sizeof(*output_codecs), &nb_output_codecs, nb_output_codecs + 1);
|
output_codecs = grow_array(output_codecs, sizeof(*output_codecs), &nb_output_codecs, nb_output_codecs + 1);
|
||||||
if (!data_stream_copy) {
|
if (!data_stream_copy) {
|
||||||
|
@ -252,7 +252,6 @@ static int a64multi_encode_frame(AVCodecContext *avctx, unsigned char *buf,
|
|||||||
int b_width;
|
int b_width;
|
||||||
|
|
||||||
int req_size;
|
int req_size;
|
||||||
int num_frames = c->mc_lifetime;
|
|
||||||
|
|
||||||
int *charmap = c->mc_charmap;
|
int *charmap = c->mc_charmap;
|
||||||
uint8_t *colram = c->mc_colram;
|
uint8_t *colram = c->mc_colram;
|
||||||
@ -280,7 +279,6 @@ static int a64multi_encode_frame(AVCodecContext *avctx, unsigned char *buf,
|
|||||||
if (!c->mc_lifetime) return 0;
|
if (!c->mc_lifetime) return 0;
|
||||||
/* no more frames in queue, prepare to flush remaining frames */
|
/* no more frames in queue, prepare to flush remaining frames */
|
||||||
if (!c->mc_frame_counter) {
|
if (!c->mc_frame_counter) {
|
||||||
num_frames = c->mc_lifetime;
|
|
||||||
c->mc_lifetime = 0;
|
c->mc_lifetime = 0;
|
||||||
}
|
}
|
||||||
/* still frames in queue so limit lifetime to remaining frames */
|
/* still frames in queue so limit lifetime to remaining frames */
|
||||||
|
@ -311,7 +311,7 @@ static void encode_window_bands_info(AACEncContext *s, SingleChannelElement *sce
|
|||||||
int win, int group_len, const float lambda)
|
int win, int group_len, const float lambda)
|
||||||
{
|
{
|
||||||
BandCodingPath path[120][12];
|
BandCodingPath path[120][12];
|
||||||
int w, swb, cb, start, start2, size;
|
int w, swb, cb, start, size;
|
||||||
int i, j;
|
int i, j;
|
||||||
const int max_sfb = sce->ics.max_sfb;
|
const int max_sfb = sce->ics.max_sfb;
|
||||||
const int run_bits = sce->ics.num_windows == 1 ? 5 : 3;
|
const int run_bits = sce->ics.num_windows == 1 ? 5 : 3;
|
||||||
@ -329,7 +329,6 @@ static void encode_window_bands_info(AACEncContext *s, SingleChannelElement *sce
|
|||||||
path[0][cb].run = 0;
|
path[0][cb].run = 0;
|
||||||
}
|
}
|
||||||
for (swb = 0; swb < max_sfb; swb++) {
|
for (swb = 0; swb < max_sfb; swb++) {
|
||||||
start2 = start;
|
|
||||||
size = sce->ics.swb_sizes[swb];
|
size = sce->ics.swb_sizes[swb];
|
||||||
if (sce->zeroes[win*16 + swb]) {
|
if (sce->zeroes[win*16 + swb]) {
|
||||||
for (cb = 0; cb < 12; cb++) {
|
for (cb = 0; cb < 12; cb++) {
|
||||||
@ -413,7 +412,7 @@ static void codebook_trellis_rate(AACEncContext *s, SingleChannelElement *sce,
|
|||||||
int win, int group_len, const float lambda)
|
int win, int group_len, const float lambda)
|
||||||
{
|
{
|
||||||
BandCodingPath path[120][12];
|
BandCodingPath path[120][12];
|
||||||
int w, swb, cb, start, start2, size;
|
int w, swb, cb, start, size;
|
||||||
int i, j;
|
int i, j;
|
||||||
const int max_sfb = sce->ics.max_sfb;
|
const int max_sfb = sce->ics.max_sfb;
|
||||||
const int run_bits = sce->ics.num_windows == 1 ? 5 : 3;
|
const int run_bits = sce->ics.num_windows == 1 ? 5 : 3;
|
||||||
@ -431,7 +430,6 @@ static void codebook_trellis_rate(AACEncContext *s, SingleChannelElement *sce,
|
|||||||
path[0][cb].run = 0;
|
path[0][cb].run = 0;
|
||||||
}
|
}
|
||||||
for (swb = 0; swb < max_sfb; swb++) {
|
for (swb = 0; swb < max_sfb; swb++) {
|
||||||
start2 = start;
|
|
||||||
size = sce->ics.swb_sizes[swb];
|
size = sce->ics.swb_sizes[swb];
|
||||||
if (sce->zeroes[win*16 + swb]) {
|
if (sce->zeroes[win*16 + swb]) {
|
||||||
for (cb = 0; cb < 12; cb++) {
|
for (cb = 0; cb < 12; cb++) {
|
||||||
@ -1006,12 +1004,11 @@ static void search_for_quantizers_fast(AVCodecContext *avctx, AACEncContext *s,
|
|||||||
SingleChannelElement *sce,
|
SingleChannelElement *sce,
|
||||||
const float lambda)
|
const float lambda)
|
||||||
{
|
{
|
||||||
int start = 0, i, w, w2, g;
|
int i, w, w2, g;
|
||||||
int minq = 255;
|
int minq = 255;
|
||||||
|
|
||||||
memset(sce->sf_idx, 0, sizeof(sce->sf_idx));
|
memset(sce->sf_idx, 0, sizeof(sce->sf_idx));
|
||||||
for (w = 0; w < sce->ics.num_windows; w += sce->ics.group_len[w]) {
|
for (w = 0; w < sce->ics.num_windows; w += sce->ics.group_len[w]) {
|
||||||
start = w*128;
|
|
||||||
for (g = 0; g < sce->ics.num_swb; g++) {
|
for (g = 0; g < sce->ics.num_swb; g++) {
|
||||||
for (w2 = 0; w2 < sce->ics.group_len[w]; w2++) {
|
for (w2 = 0; w2 < sce->ics.group_len[w]; w2++) {
|
||||||
FFPsyBand *band = &s->psy.psy_bands[s->cur_channel*PSY_MAX_BANDS+(w+w2)*16+g];
|
FFPsyBand *band = &s->psy.psy_bands[s->cur_channel*PSY_MAX_BANDS+(w+w2)*16+g];
|
||||||
|
@ -109,7 +109,7 @@ void ff_celp_lp_synthesis_filterf(float *out, const float *filter_coeffs,
|
|||||||
old_out2 = out[-2];
|
old_out2 = out[-2];
|
||||||
old_out3 = out[-1];
|
old_out3 = out[-1];
|
||||||
for (n = 0; n <= buffer_length - 4; n+=4) {
|
for (n = 0; n <= buffer_length - 4; n+=4) {
|
||||||
float tmp0,tmp1,tmp2,tmp3;
|
float tmp0,tmp1,tmp2;
|
||||||
float val;
|
float val;
|
||||||
|
|
||||||
out0 = in[0];
|
out0 = in[0];
|
||||||
@ -160,7 +160,6 @@ void ff_celp_lp_synthesis_filterf(float *out, const float *filter_coeffs,
|
|||||||
tmp0 = out0;
|
tmp0 = out0;
|
||||||
tmp1 = out1;
|
tmp1 = out1;
|
||||||
tmp2 = out2;
|
tmp2 = out2;
|
||||||
tmp3 = out3;
|
|
||||||
|
|
||||||
out3 -= a * tmp2;
|
out3 -= a * tmp2;
|
||||||
out2 -= a * tmp1;
|
out2 -= a * tmp1;
|
||||||
|
@ -1535,8 +1535,6 @@ static void dca_exss_parse_header(DCAContext *s)
|
|||||||
{
|
{
|
||||||
int ss_index;
|
int ss_index;
|
||||||
int blownup;
|
int blownup;
|
||||||
int header_size;
|
|
||||||
int hd_size;
|
|
||||||
int num_audiop = 1;
|
int num_audiop = 1;
|
||||||
int num_assets = 1;
|
int num_assets = 1;
|
||||||
int active_ss_mask[8];
|
int active_ss_mask[8];
|
||||||
@ -1549,8 +1547,8 @@ static void dca_exss_parse_header(DCAContext *s)
|
|||||||
ss_index = get_bits(&s->gb, 2);
|
ss_index = get_bits(&s->gb, 2);
|
||||||
|
|
||||||
blownup = get_bits1(&s->gb);
|
blownup = get_bits1(&s->gb);
|
||||||
header_size = get_bits(&s->gb, 8 + 4 * blownup) + 1;
|
skip_bits(&s->gb, 8 + 4 * blownup); // header_size
|
||||||
hd_size = get_bits_long(&s->gb, 16 + 4 * blownup) + 1;
|
skip_bits(&s->gb, 16 + 4 * blownup); // hd_size
|
||||||
|
|
||||||
s->static_fields = get_bits1(&s->gb);
|
s->static_fields = get_bits1(&s->gb);
|
||||||
if (s->static_fields) {
|
if (s->static_fields) {
|
||||||
|
@ -245,11 +245,11 @@ static int parse_source_parameters(AVCodecContext *avctx, GetBitContext *gb,
|
|||||||
int ff_dirac_parse_sequence_header(AVCodecContext *avctx, GetBitContext *gb,
|
int ff_dirac_parse_sequence_header(AVCodecContext *avctx, GetBitContext *gb,
|
||||||
dirac_source_params *source)
|
dirac_source_params *source)
|
||||||
{
|
{
|
||||||
unsigned version_major, version_minor;
|
unsigned version_major;
|
||||||
unsigned video_format, picture_coding_mode;
|
unsigned video_format, picture_coding_mode;
|
||||||
|
|
||||||
version_major = svq3_get_ue_golomb(gb);
|
version_major = svq3_get_ue_golomb(gb);
|
||||||
version_minor = svq3_get_ue_golomb(gb);
|
svq3_get_ue_golomb(gb); /* version_minor */
|
||||||
avctx->profile = svq3_get_ue_golomb(gb);
|
avctx->profile = svq3_get_ue_golomb(gb);
|
||||||
avctx->level = svq3_get_ue_golomb(gb);
|
avctx->level = svq3_get_ue_golomb(gb);
|
||||||
video_format = svq3_get_ue_golomb(gb);
|
video_format = svq3_get_ue_golomb(gb);
|
||||||
|
@ -152,7 +152,7 @@ int ff_h263_decode_mba(MpegEncContext *s)
|
|||||||
*/
|
*/
|
||||||
static int h263_decode_gob_header(MpegEncContext *s)
|
static int h263_decode_gob_header(MpegEncContext *s)
|
||||||
{
|
{
|
||||||
unsigned int val, gfid, gob_number;
|
unsigned int val, gob_number;
|
||||||
int left;
|
int left;
|
||||||
|
|
||||||
/* Check for GOB Start Code */
|
/* Check for GOB Start Code */
|
||||||
@ -183,12 +183,12 @@ static int h263_decode_gob_header(MpegEncContext *s)
|
|||||||
s->qscale = get_bits(&s->gb, 5); /* SQUANT */
|
s->qscale = get_bits(&s->gb, 5); /* SQUANT */
|
||||||
if(get_bits1(&s->gb)==0)
|
if(get_bits1(&s->gb)==0)
|
||||||
return -1;
|
return -1;
|
||||||
gfid = get_bits(&s->gb, 2); /* GFID */
|
skip_bits(&s->gb, 2); /* GFID */
|
||||||
}else{
|
}else{
|
||||||
gob_number = get_bits(&s->gb, 5); /* GN */
|
gob_number = get_bits(&s->gb, 5); /* GN */
|
||||||
s->mb_x= 0;
|
s->mb_x= 0;
|
||||||
s->mb_y= s->gob_index* gob_number;
|
s->mb_y= s->gob_index* gob_number;
|
||||||
gfid = get_bits(&s->gb, 2); /* GFID */
|
skip_bits(&s->gb, 2); /* GFID */
|
||||||
s->qscale = get_bits(&s->gb, 5); /* GQUANT */
|
s->qscale = get_bits(&s->gb, 5); /* GQUANT */
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2128,14 +2128,13 @@ static void mpeg_decode_gop(AVCodecContext *avctx,
|
|||||||
Mpeg1Context *s1 = avctx->priv_data;
|
Mpeg1Context *s1 = avctx->priv_data;
|
||||||
MpegEncContext *s = &s1->mpeg_enc_ctx;
|
MpegEncContext *s = &s1->mpeg_enc_ctx;
|
||||||
|
|
||||||
int drop_frame_flag;
|
|
||||||
int time_code_hours, time_code_minutes;
|
int time_code_hours, time_code_minutes;
|
||||||
int time_code_seconds, time_code_pictures;
|
int time_code_seconds, time_code_pictures;
|
||||||
int broken_link;
|
int broken_link;
|
||||||
|
|
||||||
init_get_bits(&s->gb, buf, buf_size*8);
|
init_get_bits(&s->gb, buf, buf_size*8);
|
||||||
|
|
||||||
drop_frame_flag = get_bits1(&s->gb);
|
skip_bits1(&s->gb); /* drop_frame_flag */
|
||||||
|
|
||||||
time_code_hours=get_bits(&s->gb,5);
|
time_code_hours=get_bits(&s->gb,5);
|
||||||
time_code_minutes = get_bits(&s->gb,6);
|
time_code_minutes = get_bits(&s->gb,6);
|
||||||
|
@ -761,10 +761,9 @@ void ff_mpeg1_encode_init(MpegEncContext *s)
|
|||||||
|
|
||||||
if(mv==0) len= ff_mpeg12_mbMotionVectorTable[0][1];
|
if(mv==0) len= ff_mpeg12_mbMotionVectorTable[0][1];
|
||||||
else{
|
else{
|
||||||
int val, bit_size, range, code;
|
int val, bit_size, code;
|
||||||
|
|
||||||
bit_size = f_code - 1;
|
bit_size = f_code - 1;
|
||||||
range = 1 << bit_size;
|
|
||||||
|
|
||||||
val=mv;
|
val=mv;
|
||||||
if (val < 0)
|
if (val < 0)
|
||||||
|
@ -397,14 +397,13 @@ int mpeg4_decode_video_packet_header(MpegEncContext *s)
|
|||||||
header_extension= get_bits1(&s->gb);
|
header_extension= get_bits1(&s->gb);
|
||||||
}
|
}
|
||||||
if(header_extension){
|
if(header_extension){
|
||||||
int time_increment;
|
|
||||||
int time_incr=0;
|
int time_incr=0;
|
||||||
|
|
||||||
while (get_bits1(&s->gb) != 0)
|
while (get_bits1(&s->gb) != 0)
|
||||||
time_incr++;
|
time_incr++;
|
||||||
|
|
||||||
check_marker(&s->gb, "before time_increment in video packed header");
|
check_marker(&s->gb, "before time_increment in video packed header");
|
||||||
time_increment= get_bits(&s->gb, s->time_increment_bits);
|
skip_bits(&s->gb, s->time_increment_bits); /* time_increment */
|
||||||
check_marker(&s->gb, "before vop_coding_type in video packed header");
|
check_marker(&s->gb, "before vop_coding_type in video packed header");
|
||||||
|
|
||||||
skip_bits(&s->gb, 2); /* vop coding type */
|
skip_bits(&s->gb, 2); /* vop coding type */
|
||||||
@ -1801,16 +1800,14 @@ no_cplx_est:
|
|||||||
|
|
||||||
if (s->scalability) {
|
if (s->scalability) {
|
||||||
GetBitContext bak= *gb;
|
GetBitContext bak= *gb;
|
||||||
int ref_layer_id;
|
|
||||||
int ref_layer_sampling_dir;
|
|
||||||
int h_sampling_factor_n;
|
int h_sampling_factor_n;
|
||||||
int h_sampling_factor_m;
|
int h_sampling_factor_m;
|
||||||
int v_sampling_factor_n;
|
int v_sampling_factor_n;
|
||||||
int v_sampling_factor_m;
|
int v_sampling_factor_m;
|
||||||
|
|
||||||
s->hierachy_type= get_bits1(gb);
|
s->hierachy_type= get_bits1(gb);
|
||||||
ref_layer_id= get_bits(gb, 4);
|
skip_bits(gb, 4); /* ref_layer_id */
|
||||||
ref_layer_sampling_dir= get_bits1(gb);
|
skip_bits1(gb); /* ref_layer_sampling_dir */
|
||||||
h_sampling_factor_n= get_bits(gb, 5);
|
h_sampling_factor_n= get_bits(gb, 5);
|
||||||
h_sampling_factor_m= get_bits(gb, 5);
|
h_sampling_factor_m= get_bits(gb, 5);
|
||||||
v_sampling_factor_n= get_bits(gb, 5);
|
v_sampling_factor_n= get_bits(gb, 5);
|
||||||
@ -1989,15 +1986,13 @@ static int decode_vop_header(MpegEncContext *s, GetBitContext *gb){
|
|||||||
|
|
||||||
if (s->shape != RECT_SHAPE) {
|
if (s->shape != RECT_SHAPE) {
|
||||||
if (s->vol_sprite_usage != 1 || s->pict_type != AV_PICTURE_TYPE_I) {
|
if (s->vol_sprite_usage != 1 || s->pict_type != AV_PICTURE_TYPE_I) {
|
||||||
int width, height, hor_spat_ref, ver_spat_ref;
|
skip_bits(gb, 13); /* width */
|
||||||
|
|
||||||
width = get_bits(gb, 13);
|
|
||||||
skip_bits1(gb); /* marker */
|
skip_bits1(gb); /* marker */
|
||||||
height = get_bits(gb, 13);
|
skip_bits(gb, 13); /* height */
|
||||||
skip_bits1(gb); /* marker */
|
skip_bits1(gb); /* marker */
|
||||||
hor_spat_ref = get_bits(gb, 13); /* hor_spat_ref */
|
skip_bits(gb, 13); /* hor_spat_ref */
|
||||||
skip_bits1(gb); /* marker */
|
skip_bits1(gb); /* marker */
|
||||||
ver_spat_ref = get_bits(gb, 13); /* ver_spat_ref */
|
skip_bits(gb, 13); /* ver_spat_ref */
|
||||||
}
|
}
|
||||||
skip_bits1(gb); /* change_CR_disable */
|
skip_bits1(gb); /* change_CR_disable */
|
||||||
|
|
||||||
|
@ -32,7 +32,7 @@ static void mpegvideo_extract_headers(AVCodecParserContext *s,
|
|||||||
uint32_t start_code;
|
uint32_t start_code;
|
||||||
int frame_rate_index, ext_type, bytes_left;
|
int frame_rate_index, ext_type, bytes_left;
|
||||||
int frame_rate_ext_n, frame_rate_ext_d;
|
int frame_rate_ext_n, frame_rate_ext_d;
|
||||||
int picture_structure, top_field_first, repeat_first_field, progressive_frame;
|
int top_field_first, repeat_first_field, progressive_frame;
|
||||||
int horiz_size_ext, vert_size_ext, bit_rate_ext;
|
int horiz_size_ext, vert_size_ext, bit_rate_ext;
|
||||||
int did_set_size=0;
|
int did_set_size=0;
|
||||||
//FIXME replace the crap with get_bits()
|
//FIXME replace the crap with get_bits()
|
||||||
@ -91,7 +91,6 @@ static void mpegvideo_extract_headers(AVCodecParserContext *s,
|
|||||||
break;
|
break;
|
||||||
case 0x8: /* picture coding extension */
|
case 0x8: /* picture coding extension */
|
||||||
if (bytes_left >= 5) {
|
if (bytes_left >= 5) {
|
||||||
picture_structure = buf[2]&3;
|
|
||||||
top_field_first = buf[3] & (1 << 7);
|
top_field_first = buf[3] & (1 << 7);
|
||||||
repeat_first_field = buf[3] & (1 << 1);
|
repeat_first_field = buf[3] & (1 << 1);
|
||||||
progressive_frame = buf[4] & (1 << 7);
|
progressive_frame = buf[4] & (1 << 7);
|
||||||
|
@ -1528,9 +1528,7 @@ int msmpeg4_decode_ext_header(MpegEncContext * s, int buf_size)
|
|||||||
/* the alt_bitstream reader could read over the end so we need to check it */
|
/* the alt_bitstream reader could read over the end so we need to check it */
|
||||||
if(left>=length && left<length+8)
|
if(left>=length && left<length+8)
|
||||||
{
|
{
|
||||||
int fps;
|
skip_bits(&s->gb, 5); /* fps */
|
||||||
|
|
||||||
fps= get_bits(&s->gb, 5);
|
|
||||||
s->bit_rate= get_bits(&s->gb, 11)*1024;
|
s->bit_rate= get_bits(&s->gb, 11)*1024;
|
||||||
if(s->msmpeg4_version>=3)
|
if(s->msmpeg4_version>=3)
|
||||||
s->flipflop_rounding= get_bits1(&s->gb);
|
s->flipflop_rounding= get_bits1(&s->gb);
|
||||||
|
@ -397,7 +397,7 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
AVFrame *p;
|
AVFrame *p;
|
||||||
uint8_t *crow_buf_base = NULL;
|
uint8_t *crow_buf_base = NULL;
|
||||||
uint32_t tag, length;
|
uint32_t tag, length;
|
||||||
int ret, crc;
|
int ret;
|
||||||
|
|
||||||
FFSWAP(AVFrame *, s->current_picture, s->last_picture);
|
FFSWAP(AVFrame *, s->current_picture, s->last_picture);
|
||||||
avctx->coded_frame= s->current_picture;
|
avctx->coded_frame= s->current_picture;
|
||||||
@ -451,7 +451,7 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
s->compression_type = *s->bytestream++;
|
s->compression_type = *s->bytestream++;
|
||||||
s->filter_type = *s->bytestream++;
|
s->filter_type = *s->bytestream++;
|
||||||
s->interlace_type = *s->bytestream++;
|
s->interlace_type = *s->bytestream++;
|
||||||
crc = bytestream_get_be32(&s->bytestream);
|
s->bytestream += 4; /* crc */
|
||||||
s->state |= PNG_IHDR;
|
s->state |= PNG_IHDR;
|
||||||
av_dlog(avctx, "width=%d height=%d depth=%d color_type=%d compression_type=%d filter_type=%d interlace_type=%d\n",
|
av_dlog(avctx, "width=%d height=%d depth=%d color_type=%d compression_type=%d filter_type=%d interlace_type=%d\n",
|
||||||
s->width, s->height, s->bit_depth, s->color_type,
|
s->width, s->height, s->bit_depth, s->color_type,
|
||||||
@ -547,8 +547,7 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
s->state |= PNG_IDAT;
|
s->state |= PNG_IDAT;
|
||||||
if (png_decode_idat(s, length) < 0)
|
if (png_decode_idat(s, length) < 0)
|
||||||
goto fail;
|
goto fail;
|
||||||
/* skip crc */
|
s->bytestream += 4; /* crc */
|
||||||
crc = bytestream_get_be32(&s->bytestream);
|
|
||||||
break;
|
break;
|
||||||
case MKTAG('P', 'L', 'T', 'E'):
|
case MKTAG('P', 'L', 'T', 'E'):
|
||||||
{
|
{
|
||||||
@ -568,7 +567,7 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
s->palette[i] = (0xff << 24);
|
s->palette[i] = (0xff << 24);
|
||||||
}
|
}
|
||||||
s->state |= PNG_PLTE;
|
s->state |= PNG_PLTE;
|
||||||
crc = bytestream_get_be32(&s->bytestream);
|
s->bytestream += 4; /* crc */
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case MKTAG('t', 'R', 'N', 'S'):
|
case MKTAG('t', 'R', 'N', 'S'):
|
||||||
@ -584,13 +583,13 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
v = *s->bytestream++;
|
v = *s->bytestream++;
|
||||||
s->palette[i] = (s->palette[i] & 0x00ffffff) | (v << 24);
|
s->palette[i] = (s->palette[i] & 0x00ffffff) | (v << 24);
|
||||||
}
|
}
|
||||||
crc = bytestream_get_be32(&s->bytestream);
|
s->bytestream += 4; /* crc */
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case MKTAG('I', 'E', 'N', 'D'):
|
case MKTAG('I', 'E', 'N', 'D'):
|
||||||
if (!(s->state & PNG_ALLIMAGE))
|
if (!(s->state & PNG_ALLIMAGE))
|
||||||
goto fail;
|
goto fail;
|
||||||
crc = bytestream_get_be32(&s->bytestream);
|
s->bytestream += 4; /* crc */
|
||||||
goto exit_loop;
|
goto exit_loop;
|
||||||
default:
|
default:
|
||||||
/* skip tag */
|
/* skip tag */
|
||||||
|
@ -235,7 +235,7 @@ int rv_decode_dc(MpegEncContext *s, int n)
|
|||||||
/* read RV 1.0 compatible frame header */
|
/* read RV 1.0 compatible frame header */
|
||||||
static int rv10_decode_picture_header(MpegEncContext *s)
|
static int rv10_decode_picture_header(MpegEncContext *s)
|
||||||
{
|
{
|
||||||
int mb_count, pb_frame, marker, unk, mb_xy;
|
int mb_count, pb_frame, marker, mb_xy;
|
||||||
|
|
||||||
marker = get_bits1(&s->gb);
|
marker = get_bits1(&s->gb);
|
||||||
|
|
||||||
@ -282,7 +282,7 @@ static int rv10_decode_picture_header(MpegEncContext *s)
|
|||||||
s->mb_y = 0;
|
s->mb_y = 0;
|
||||||
mb_count = s->mb_width * s->mb_height;
|
mb_count = s->mb_width * s->mb_height;
|
||||||
}
|
}
|
||||||
unk= get_bits(&s->gb, 3); /* ignored */
|
skip_bits(&s->gb, 3); /* ignored */
|
||||||
s->f_code = 1;
|
s->f_code = 1;
|
||||||
s->unrestricted_mv = 1;
|
s->unrestricted_mv = 1;
|
||||||
|
|
||||||
|
@ -29,7 +29,7 @@ static int s302m_parse_frame_header(AVCodecContext *avctx, const uint8_t *buf,
|
|||||||
int buf_size)
|
int buf_size)
|
||||||
{
|
{
|
||||||
uint32_t h;
|
uint32_t h;
|
||||||
int frame_size, channels, id, bits;
|
int frame_size, channels, bits;
|
||||||
|
|
||||||
if (buf_size <= AES3_HEADER_LEN) {
|
if (buf_size <= AES3_HEADER_LEN) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "frame is too short\n");
|
av_log(avctx, AV_LOG_ERROR, "frame is too short\n");
|
||||||
@ -48,7 +48,6 @@ static int s302m_parse_frame_header(AVCodecContext *avctx, const uint8_t *buf,
|
|||||||
h = AV_RB32(buf);
|
h = AV_RB32(buf);
|
||||||
frame_size = (h >> 16) & 0xffff;
|
frame_size = (h >> 16) & 0xffff;
|
||||||
channels = ((h >> 14) & 0x0003) * 2 + 2;
|
channels = ((h >> 14) & 0x0003) * 2 + 2;
|
||||||
id = (h >> 6) & 0x00ff;
|
|
||||||
bits = ((h >> 4) & 0x0003) * 4 + 16;
|
bits = ((h >> 4) & 0x0003) * 4 + 16;
|
||||||
|
|
||||||
if (AES3_HEADER_LEN + frame_size != buf_size || bits > 24) {
|
if (AES3_HEADER_LEN + frame_size != buf_size || bits > 24) {
|
||||||
|
@ -196,7 +196,6 @@ static int decode_wave_header(AVCodecContext *avctx, uint8_t *header, int header
|
|||||||
{
|
{
|
||||||
GetBitContext hb;
|
GetBitContext hb;
|
||||||
int len;
|
int len;
|
||||||
int chunk_size;
|
|
||||||
short wave_format;
|
short wave_format;
|
||||||
|
|
||||||
init_get_bits(&hb, header, header_size*8);
|
init_get_bits(&hb, header, header_size*8);
|
||||||
@ -205,7 +204,7 @@ static int decode_wave_header(AVCodecContext *avctx, uint8_t *header, int header
|
|||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
chunk_size = get_le32(&hb);
|
skip_bits_long(&hb, 32); /* chunk_size */
|
||||||
|
|
||||||
if (get_le32(&hb) != MKTAG('W','A','V','E')) {
|
if (get_le32(&hb) != MKTAG('W','A','V','E')) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "missing WAVE tag\n");
|
av_log(avctx, AV_LOG_ERROR, "missing WAVE tag\n");
|
||||||
|
@ -38,15 +38,12 @@ static int sp5x_decode_frame(AVCodecContext *avctx,
|
|||||||
int buf_size = avpkt->size;
|
int buf_size = avpkt->size;
|
||||||
AVPacket avpkt_recoded;
|
AVPacket avpkt_recoded;
|
||||||
const int qscale = 5;
|
const int qscale = 5;
|
||||||
const uint8_t *buf_ptr;
|
|
||||||
uint8_t *recoded;
|
uint8_t *recoded;
|
||||||
int i = 0, j = 0;
|
int i = 0, j = 0;
|
||||||
|
|
||||||
if (!avctx->width || !avctx->height)
|
if (!avctx->width || !avctx->height)
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
buf_ptr = buf;
|
|
||||||
|
|
||||||
recoded = av_mallocz(buf_size + 1024);
|
recoded = av_mallocz(buf_size + 1024);
|
||||||
if (!recoded)
|
if (!recoded)
|
||||||
return -1;
|
return -1;
|
||||||
|
@ -554,9 +554,8 @@ static void svq1_parse_string (GetBitContext *bitbuf, uint8_t *out) {
|
|||||||
|
|
||||||
static int svq1_decode_frame_header (GetBitContext *bitbuf,MpegEncContext *s) {
|
static int svq1_decode_frame_header (GetBitContext *bitbuf,MpegEncContext *s) {
|
||||||
int frame_size_code;
|
int frame_size_code;
|
||||||
int temporal_reference;
|
|
||||||
|
|
||||||
temporal_reference = get_bits (bitbuf, 8);
|
skip_bits(bitbuf, 8); /* temporal_reference */
|
||||||
|
|
||||||
/* frame type */
|
/* frame type */
|
||||||
s->pict_type= get_bits (bitbuf, 2)+1;
|
s->pict_type= get_bits (bitbuf, 2)+1;
|
||||||
|
@ -108,18 +108,18 @@ static int decode_frame(AVCodecContext *avctx,
|
|||||||
AVFrame * const p= (AVFrame*)&s->picture;
|
AVFrame * const p= (AVFrame*)&s->picture;
|
||||||
uint8_t *dst;
|
uint8_t *dst;
|
||||||
int stride;
|
int stride;
|
||||||
int idlen, pal, compr, x, y, w, h, bpp, flags;
|
int idlen, compr, y, w, h, bpp, flags;
|
||||||
int first_clr, colors, csize;
|
int first_clr, colors, csize;
|
||||||
|
|
||||||
/* parse image header */
|
/* parse image header */
|
||||||
CHECK_BUFFER_SIZE(buf, buf_end, 18, "header");
|
CHECK_BUFFER_SIZE(buf, buf_end, 18, "header");
|
||||||
idlen = *buf++;
|
idlen = *buf++;
|
||||||
pal = *buf++;
|
buf++; /* pal */
|
||||||
compr = *buf++;
|
compr = *buf++;
|
||||||
first_clr = AV_RL16(buf); buf += 2;
|
first_clr = AV_RL16(buf); buf += 2;
|
||||||
colors = AV_RL16(buf); buf += 2;
|
colors = AV_RL16(buf); buf += 2;
|
||||||
csize = *buf++;
|
csize = *buf++;
|
||||||
x = AV_RL16(buf); buf += 2;
|
buf += 2; /* x */
|
||||||
y = AV_RL16(buf); buf += 2;
|
y = AV_RL16(buf); buf += 2;
|
||||||
w = AV_RL16(buf); buf += 2;
|
w = AV_RL16(buf); buf += 2;
|
||||||
h = AV_RL16(buf); buf += 2;
|
h = AV_RL16(buf); buf += 2;
|
||||||
|
@ -201,7 +201,6 @@ static inline int tm2_read_header(TM2Context *ctx, const uint8_t *buf)
|
|||||||
{
|
{
|
||||||
uint32_t magic;
|
uint32_t magic;
|
||||||
const uint8_t *obuf;
|
const uint8_t *obuf;
|
||||||
int length;
|
|
||||||
|
|
||||||
obuf = buf;
|
obuf = buf;
|
||||||
|
|
||||||
@ -212,19 +211,6 @@ static inline int tm2_read_header(TM2Context *ctx, const uint8_t *buf)
|
|||||||
/* av_log (ctx->avctx, AV_LOG_ERROR, "TM2 old header: not implemented (yet)\n"); */
|
/* av_log (ctx->avctx, AV_LOG_ERROR, "TM2 old header: not implemented (yet)\n"); */
|
||||||
return 40;
|
return 40;
|
||||||
} else if(magic == 0x00000101) { /* new header */
|
} else if(magic == 0x00000101) { /* new header */
|
||||||
int w, h, size, flags, xr, yr;
|
|
||||||
|
|
||||||
length = AV_RL32(buf);
|
|
||||||
buf += 4;
|
|
||||||
|
|
||||||
init_get_bits(&ctx->gb, buf, 32 * 8);
|
|
||||||
size = get_bits_long(&ctx->gb, 31);
|
|
||||||
h = get_bits(&ctx->gb, 15);
|
|
||||||
w = get_bits(&ctx->gb, 15);
|
|
||||||
flags = get_bits_long(&ctx->gb, 31);
|
|
||||||
yr = get_bits(&ctx->gb, 9);
|
|
||||||
xr = get_bits(&ctx->gb, 9);
|
|
||||||
|
|
||||||
return 40;
|
return 40;
|
||||||
} else {
|
} else {
|
||||||
av_log (ctx->avctx, AV_LOG_ERROR, "Not a TM2 header: 0x%08X\n", magic);
|
av_log (ctx->avctx, AV_LOG_ERROR, "Not a TM2 header: 0x%08X\n", magic);
|
||||||
|
@ -75,7 +75,6 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac
|
|||||||
int buf_size = avpkt->size;
|
int buf_size = avpkt->size;
|
||||||
CamtasiaContext * const c = avctx->priv_data;
|
CamtasiaContext * const c = avctx->priv_data;
|
||||||
const unsigned char *encoded = buf;
|
const unsigned char *encoded = buf;
|
||||||
unsigned char *outptr;
|
|
||||||
int zret; // Zlib return code
|
int zret; // Zlib return code
|
||||||
int len = buf_size;
|
int len = buf_size;
|
||||||
|
|
||||||
@ -89,8 +88,6 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac
|
|||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
outptr = c->pic.data[0]; // Output image pointer
|
|
||||||
|
|
||||||
zret = inflateReset(&(c->zstream));
|
zret = inflateReset(&(c->zstream));
|
||||||
if (zret != Z_OK) {
|
if (zret != Z_OK) {
|
||||||
av_log(avctx, AV_LOG_ERROR, "Inflate reset error: %d\n", zret);
|
av_log(avctx, AV_LOG_ERROR, "Inflate reset error: %d\n", zret);
|
||||||
|
@ -199,7 +199,6 @@ static void vmd_decode(VmdVideoContext *s)
|
|||||||
|
|
||||||
int frame_x, frame_y;
|
int frame_x, frame_y;
|
||||||
int frame_width, frame_height;
|
int frame_width, frame_height;
|
||||||
int dp_size;
|
|
||||||
|
|
||||||
frame_x = AV_RL16(&s->buf[6]);
|
frame_x = AV_RL16(&s->buf[6]);
|
||||||
frame_y = AV_RL16(&s->buf[8]);
|
frame_y = AV_RL16(&s->buf[8]);
|
||||||
@ -247,7 +246,6 @@ static void vmd_decode(VmdVideoContext *s)
|
|||||||
}
|
}
|
||||||
|
|
||||||
dp = &s->frame.data[0][frame_y * s->frame.linesize[0] + frame_x];
|
dp = &s->frame.data[0][frame_y * s->frame.linesize[0] + frame_x];
|
||||||
dp_size = s->frame.linesize[0] * s->avctx->height;
|
|
||||||
pp = &s->prev_frame.data[0][frame_y * s->prev_frame.linesize[0] + frame_x];
|
pp = &s->prev_frame.data[0][frame_y * s->prev_frame.linesize[0] + frame_x];
|
||||||
switch (meth) {
|
switch (meth) {
|
||||||
case 1:
|
case 1:
|
||||||
|
@ -181,7 +181,7 @@ static int encode_frame(AVCodecContext *avctx, uint8_t *buf, int buf_size, void
|
|||||||
int x, y, bh2, bw2, xored;
|
int x, y, bh2, bw2, xored;
|
||||||
uint8_t *tsrc, *tprev;
|
uint8_t *tsrc, *tprev;
|
||||||
uint8_t *mv;
|
uint8_t *mv;
|
||||||
int mx, my, bv;
|
int mx, my;
|
||||||
|
|
||||||
bw = (avctx->width + ZMBV_BLOCK - 1) / ZMBV_BLOCK;
|
bw = (avctx->width + ZMBV_BLOCK - 1) / ZMBV_BLOCK;
|
||||||
bh = (avctx->height + ZMBV_BLOCK - 1) / ZMBV_BLOCK;
|
bh = (avctx->height + ZMBV_BLOCK - 1) / ZMBV_BLOCK;
|
||||||
@ -197,7 +197,7 @@ static int encode_frame(AVCodecContext *avctx, uint8_t *buf, int buf_size, void
|
|||||||
tsrc = src + x;
|
tsrc = src + x;
|
||||||
tprev = prev + x;
|
tprev = prev + x;
|
||||||
|
|
||||||
bv = zmbv_me(c, tsrc, p->linesize[0], tprev, c->pstride, x, y, &mx, &my, &xored);
|
zmbv_me(c, tsrc, p->linesize[0], tprev, c->pstride, x, y, &mx, &my, &xored);
|
||||||
mv[0] = (mx << 1) | !!xored;
|
mv[0] = (mx << 1) | !!xored;
|
||||||
mv[1] = my << 1;
|
mv[1] = my << 1;
|
||||||
tprev += mx + my * c->pstride;
|
tprev += mx + my * c->pstride;
|
||||||
|
@ -153,7 +153,7 @@ static void end_frame(AVFilterLink *inlink)
|
|||||||
AVFilterBufferRef *inpicref = inlink->cur_buf;
|
AVFilterBufferRef *inpicref = inlink->cur_buf;
|
||||||
AVFilterBufferRef *outpicref = outlink->out_buf;
|
AVFilterBufferRef *outpicref = outlink->out_buf;
|
||||||
|
|
||||||
int h, w, plane, line_step, line_size, line;
|
int h, plane, line_step, line_size, line;
|
||||||
uint8_t *cpy_src, *cpy_dst;
|
uint8_t *cpy_src, *cpy_dst;
|
||||||
|
|
||||||
if ( inpicref->video->interlaced
|
if ( inpicref->video->interlaced
|
||||||
@ -162,7 +162,6 @@ static void end_frame(AVFilterLink *inlink)
|
|||||||
"picture will move %s one line\n",
|
"picture will move %s one line\n",
|
||||||
fieldorder->dst_tff ? "up" : "down");
|
fieldorder->dst_tff ? "up" : "down");
|
||||||
h = inpicref->video->h;
|
h = inpicref->video->h;
|
||||||
w = inpicref->video->w;
|
|
||||||
for (plane = 0; plane < 4 && inpicref->data[plane]; plane++) {
|
for (plane = 0; plane < 4 && inpicref->data[plane]; plane++) {
|
||||||
line_step = inpicref->linesize[plane];
|
line_step = inpicref->linesize[plane];
|
||||||
line_size = fieldorder->line_size[plane];
|
line_size = fieldorder->line_size[plane];
|
||||||
|
@ -246,7 +246,7 @@ static int fourxm_read_packet(AVFormatContext *s,
|
|||||||
FourxmDemuxContext *fourxm = s->priv_data;
|
FourxmDemuxContext *fourxm = s->priv_data;
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
unsigned int fourcc_tag;
|
unsigned int fourcc_tag;
|
||||||
unsigned int size, out_size;
|
unsigned int size;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
unsigned int track_number;
|
unsigned int track_number;
|
||||||
int packet_read = 0;
|
int packet_read = 0;
|
||||||
@ -295,7 +295,7 @@ static int fourxm_read_packet(AVFormatContext *s,
|
|||||||
|
|
||||||
case snd__TAG:
|
case snd__TAG:
|
||||||
track_number = avio_rl32(pb);
|
track_number = avio_rl32(pb);
|
||||||
out_size= avio_rl32(pb);
|
avio_skip(pb, 4);
|
||||||
size-=8;
|
size-=8;
|
||||||
|
|
||||||
if (track_number < fourxm->track_count && fourxm->tracks[track_number].channels>0) {
|
if (track_number < fourxm->track_count && fourxm->tracks[track_number].channels>0) {
|
||||||
|
@ -35,11 +35,11 @@ static int ape_tag_read_field(AVFormatContext *s)
|
|||||||
{
|
{
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
uint8_t key[1024], *value;
|
uint8_t key[1024], *value;
|
||||||
uint32_t size, flags;
|
uint32_t size;
|
||||||
int i, c;
|
int i, c;
|
||||||
|
|
||||||
size = avio_rl32(pb); /* field size */
|
size = avio_rl32(pb); /* field size */
|
||||||
flags = avio_rl32(pb); /* field flags */
|
avio_skip(pb, 4); /* field flags */
|
||||||
for (i = 0; i < sizeof(key) - 1; i++) {
|
for (i = 0; i < sizeof(key) - 1; i++) {
|
||||||
c = avio_r8(pb);
|
c = avio_r8(pb);
|
||||||
if (c < 0x20 || c > 0x7E)
|
if (c < 0x20 || c > 0x7E)
|
||||||
|
@ -216,7 +216,6 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
|
|||||||
ff_asf_guid g;
|
ff_asf_guid g;
|
||||||
enum AVMediaType type;
|
enum AVMediaType type;
|
||||||
int type_specific_size, sizeX;
|
int type_specific_size, sizeX;
|
||||||
uint64_t total_size;
|
|
||||||
unsigned int tag1;
|
unsigned int tag1;
|
||||||
int64_t pos1, pos2, start_time;
|
int64_t pos1, pos2, start_time;
|
||||||
int test_for_ext_stream_audio, is_dvr_ms_audio=0;
|
int test_for_ext_stream_audio, is_dvr_ms_audio=0;
|
||||||
@ -264,7 +263,7 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
|
|||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
ff_get_guid(pb, &g);
|
ff_get_guid(pb, &g);
|
||||||
total_size = avio_rl64(pb);
|
avio_skip(pb, 8); /* total_size */
|
||||||
type_specific_size = avio_rl32(pb);
|
type_specific_size = avio_rl32(pb);
|
||||||
avio_rl32(pb);
|
avio_rl32(pb);
|
||||||
st->id = avio_rl16(pb) & 0x7f; /* stream id */
|
st->id = avio_rl16(pb) & 0x7f; /* stream id */
|
||||||
@ -401,7 +400,7 @@ static int asf_read_ext_stream_properties(AVFormatContext *s, int64_t size)
|
|||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
ff_asf_guid g;
|
ff_asf_guid g;
|
||||||
int ext_len, payload_ext_ct, stream_ct, i;
|
int ext_len, payload_ext_ct, stream_ct, i;
|
||||||
uint32_t ext_d, leak_rate, stream_num;
|
uint32_t leak_rate, stream_num;
|
||||||
unsigned int stream_languageid_index;
|
unsigned int stream_languageid_index;
|
||||||
|
|
||||||
avio_rl64(pb); // starttime
|
avio_rl64(pb); // starttime
|
||||||
@ -435,7 +434,7 @@ static int asf_read_ext_stream_properties(AVFormatContext *s, int64_t size)
|
|||||||
|
|
||||||
for (i=0; i<payload_ext_ct; i++){
|
for (i=0; i<payload_ext_ct; i++){
|
||||||
ff_get_guid(pb, &g);
|
ff_get_guid(pb, &g);
|
||||||
ext_d=avio_rl16(pb);
|
avio_skip(pb, 2);
|
||||||
ext_len=avio_rl32(pb);
|
ext_len=avio_rl32(pb);
|
||||||
avio_skip(pb, ext_len);
|
avio_skip(pb, ext_len);
|
||||||
}
|
}
|
||||||
@ -519,7 +518,7 @@ static int asf_read_metadata(AVFormatContext *s, int64_t size)
|
|||||||
{
|
{
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
ASFContext *asf = s->priv_data;
|
ASFContext *asf = s->priv_data;
|
||||||
int n, stream_num, name_len, value_len, value_type, value_num;
|
int n, stream_num, name_len, value_len, value_num;
|
||||||
int ret, i;
|
int ret, i;
|
||||||
n = avio_rl16(pb);
|
n = avio_rl16(pb);
|
||||||
|
|
||||||
@ -529,7 +528,7 @@ static int asf_read_metadata(AVFormatContext *s, int64_t size)
|
|||||||
avio_rl16(pb); //lang_list_index
|
avio_rl16(pb); //lang_list_index
|
||||||
stream_num= avio_rl16(pb);
|
stream_num= avio_rl16(pb);
|
||||||
name_len= avio_rl16(pb);
|
name_len= avio_rl16(pb);
|
||||||
value_type= avio_rl16(pb);
|
avio_skip(pb, 2); /* value_type */
|
||||||
value_len= avio_rl32(pb);
|
value_len= avio_rl32(pb);
|
||||||
|
|
||||||
if ((ret = avio_get_str16le(pb, name_len, name, sizeof(name))) < name_len)
|
if ((ret = avio_get_str16le(pb, name_len, name, sizeof(name))) < name_len)
|
||||||
@ -634,10 +633,8 @@ static int asf_read_header(AVFormatContext *s, AVFormatParameters *ap)
|
|||||||
// if so the next iteration will pick it up
|
// if so the next iteration will pick it up
|
||||||
continue;
|
continue;
|
||||||
} else if (!ff_guidcmp(&g, &ff_asf_head1_guid)) {
|
} else if (!ff_guidcmp(&g, &ff_asf_head1_guid)) {
|
||||||
int v1, v2;
|
|
||||||
ff_get_guid(pb, &g);
|
ff_get_guid(pb, &g);
|
||||||
v1 = avio_rl32(pb);
|
avio_skip(pb, 6);
|
||||||
v2 = avio_rl16(pb);
|
|
||||||
continue;
|
continue;
|
||||||
} else if (!ff_guidcmp(&g, &ff_asf_marker_header)) {
|
} else if (!ff_guidcmp(&g, &ff_asf_marker_header)) {
|
||||||
asf_read_marker(s, gsize);
|
asf_read_marker(s, gsize);
|
||||||
@ -807,7 +804,7 @@ static int asf_read_frame_header(AVFormatContext *s, AVIOContext *pb){
|
|||||||
ASFContext *asf = s->priv_data;
|
ASFContext *asf = s->priv_data;
|
||||||
int rsize = 1;
|
int rsize = 1;
|
||||||
int num = avio_r8(pb);
|
int num = avio_r8(pb);
|
||||||
int64_t ts0, ts1;
|
int64_t ts0;
|
||||||
|
|
||||||
asf->packet_segments--;
|
asf->packet_segments--;
|
||||||
asf->packet_key_frame = num >> 7;
|
asf->packet_key_frame = num >> 7;
|
||||||
@ -830,7 +827,7 @@ static int asf_read_frame_header(AVFormatContext *s, AVIOContext *pb){
|
|||||||
// av_log(s, AV_LOG_DEBUG, "\n");
|
// av_log(s, AV_LOG_DEBUG, "\n");
|
||||||
avio_skip(pb, 10);
|
avio_skip(pb, 10);
|
||||||
ts0= avio_rl64(pb);
|
ts0= avio_rl64(pb);
|
||||||
ts1= avio_rl64(pb);
|
avio_skip(pb, 8);;
|
||||||
avio_skip(pb, 12);
|
avio_skip(pb, 12);
|
||||||
avio_rl32(pb);
|
avio_rl32(pb);
|
||||||
avio_skip(pb, asf->packet_replic_size - 8 - 38 - 4);
|
avio_skip(pb, asf->packet_replic_size - 8 - 38 - 4);
|
||||||
|
@ -337,7 +337,7 @@ static int avi_read_header(AVFormatContext *s, AVFormatParameters *ap)
|
|||||||
AVIContext *avi = s->priv_data;
|
AVIContext *avi = s->priv_data;
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
unsigned int tag, tag1, handler;
|
unsigned int tag, tag1, handler;
|
||||||
int codec_type, stream_index, frame_period, bit_rate;
|
int codec_type, stream_index, frame_period;
|
||||||
unsigned int size;
|
unsigned int size;
|
||||||
int i;
|
int i;
|
||||||
AVStream *st;
|
AVStream *st;
|
||||||
@ -407,7 +407,7 @@ static int avi_read_header(AVFormatContext *s, AVFormatParameters *ap)
|
|||||||
/* AVI header */
|
/* AVI header */
|
||||||
/* using frame_period is bad idea */
|
/* using frame_period is bad idea */
|
||||||
frame_period = avio_rl32(pb);
|
frame_period = avio_rl32(pb);
|
||||||
bit_rate = avio_rl32(pb) * 8;
|
avio_skip(pb, 4);
|
||||||
avio_rl32(pb);
|
avio_rl32(pb);
|
||||||
avi->non_interleaved |= avio_rl32(pb) & AVIF_MUSTUSEINDEX;
|
avi->non_interleaved |= avio_rl32(pb) & AVIF_MUSTUSEINDEX;
|
||||||
|
|
||||||
|
@ -295,9 +295,7 @@ static int gif_write_video(AVFormatContext *s,
|
|||||||
AVCodecContext *enc, const uint8_t *buf, int size)
|
AVCodecContext *enc, const uint8_t *buf, int size)
|
||||||
{
|
{
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
GIFContext *gif = s->priv_data;
|
|
||||||
int jiffies;
|
int jiffies;
|
||||||
int64_t delay;
|
|
||||||
|
|
||||||
/* graphic control extension block */
|
/* graphic control extension block */
|
||||||
avio_w8(pb, 0x21);
|
avio_w8(pb, 0x21);
|
||||||
@ -307,8 +305,6 @@ static int gif_write_video(AVFormatContext *s,
|
|||||||
|
|
||||||
/* 1 jiffy is 1/70 s */
|
/* 1 jiffy is 1/70 s */
|
||||||
/* the delay_time field indicates the number of jiffies - 1 */
|
/* the delay_time field indicates the number of jiffies - 1 */
|
||||||
delay = gif->file_time - gif->time;
|
|
||||||
|
|
||||||
/* XXX: should use delay, in order to be more accurate */
|
/* XXX: should use delay, in order to be more accurate */
|
||||||
/* instead of using the same rounded value each time */
|
/* instead of using the same rounded value each time */
|
||||||
/* XXX: don't even remember if I really use it for now */
|
/* XXX: don't even remember if I really use it for now */
|
||||||
|
@ -186,13 +186,13 @@ static int mmf_read_header(AVFormatContext *s,
|
|||||||
unsigned int tag;
|
unsigned int tag;
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
AVStream *st;
|
AVStream *st;
|
||||||
int64_t file_size, size;
|
int64_t size;
|
||||||
int rate, params;
|
int rate, params;
|
||||||
|
|
||||||
tag = avio_rl32(pb);
|
tag = avio_rl32(pb);
|
||||||
if (tag != MKTAG('M', 'M', 'M', 'D'))
|
if (tag != MKTAG('M', 'M', 'M', 'D'))
|
||||||
return -1;
|
return -1;
|
||||||
file_size = avio_rb32(pb);
|
avio_skip(pb, 4); /* file_size */
|
||||||
|
|
||||||
/* Skip some unused chunks that may or may not be present */
|
/* Skip some unused chunks that may or may not be present */
|
||||||
for(;; avio_skip(pb, size)) {
|
for(;; avio_skip(pb, size)) {
|
||||||
@ -263,12 +263,10 @@ static int mmf_read_packet(AVFormatContext *s,
|
|||||||
AVPacket *pkt)
|
AVPacket *pkt)
|
||||||
{
|
{
|
||||||
MMFContext *mmf = s->priv_data;
|
MMFContext *mmf = s->priv_data;
|
||||||
AVStream *st;
|
|
||||||
int ret, size;
|
int ret, size;
|
||||||
|
|
||||||
if (s->pb->eof_reached)
|
if (s->pb->eof_reached)
|
||||||
return AVERROR(EIO);
|
return AVERROR(EIO);
|
||||||
st = s->streams[0];
|
|
||||||
|
|
||||||
size = MAX_SIZE;
|
size = MAX_SIZE;
|
||||||
if(size > mmf->data_size)
|
if(size > mmf->data_size)
|
||||||
|
@ -464,21 +464,21 @@ static int mov_read_hdlr(MOVContext *c, AVIOContext *pb, MOVAtom atom)
|
|||||||
int ff_mov_read_esds(AVFormatContext *fc, AVIOContext *pb, MOVAtom atom)
|
int ff_mov_read_esds(AVFormatContext *fc, AVIOContext *pb, MOVAtom atom)
|
||||||
{
|
{
|
||||||
AVStream *st;
|
AVStream *st;
|
||||||
int tag, len;
|
int tag;
|
||||||
|
|
||||||
if (fc->nb_streams < 1)
|
if (fc->nb_streams < 1)
|
||||||
return 0;
|
return 0;
|
||||||
st = fc->streams[fc->nb_streams-1];
|
st = fc->streams[fc->nb_streams-1];
|
||||||
|
|
||||||
avio_rb32(pb); /* version + flags */
|
avio_rb32(pb); /* version + flags */
|
||||||
len = ff_mp4_read_descr(fc, pb, &tag);
|
ff_mp4_read_descr(fc, pb, &tag);
|
||||||
if (tag == MP4ESDescrTag) {
|
if (tag == MP4ESDescrTag) {
|
||||||
avio_rb16(pb); /* ID */
|
avio_rb16(pb); /* ID */
|
||||||
avio_r8(pb); /* priority */
|
avio_r8(pb); /* priority */
|
||||||
} else
|
} else
|
||||||
avio_rb16(pb); /* ID */
|
avio_rb16(pb); /* ID */
|
||||||
|
|
||||||
len = ff_mp4_read_descr(fc, pb, &tag);
|
ff_mp4_read_descr(fc, pb, &tag);
|
||||||
if (tag == MP4DecConfigDescrTag)
|
if (tag == MP4DecConfigDescrTag)
|
||||||
ff_mp4_read_dec_config_descr(fc, st, pb);
|
ff_mp4_read_dec_config_descr(fc, st, pb);
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -196,8 +196,6 @@ static int ogg_read_page(AVFormatContext *s, int *str)
|
|||||||
int flags, nsegs;
|
int flags, nsegs;
|
||||||
uint64_t gp;
|
uint64_t gp;
|
||||||
uint32_t serial;
|
uint32_t serial;
|
||||||
uint32_t seq;
|
|
||||||
uint32_t crc;
|
|
||||||
int size, idx;
|
int size, idx;
|
||||||
uint8_t sync[4];
|
uint8_t sync[4];
|
||||||
int sp = 0;
|
int sp = 0;
|
||||||
@ -230,8 +228,7 @@ static int ogg_read_page(AVFormatContext *s, int *str)
|
|||||||
flags = avio_r8(bc);
|
flags = avio_r8(bc);
|
||||||
gp = avio_rl64 (bc);
|
gp = avio_rl64 (bc);
|
||||||
serial = avio_rl32 (bc);
|
serial = avio_rl32 (bc);
|
||||||
seq = avio_rl32 (bc);
|
avio_skip(bc, 8); /* seq, crc */
|
||||||
crc = avio_rl32 (bc);
|
|
||||||
nsegs = avio_r8(bc);
|
nsegs = avio_r8(bc);
|
||||||
|
|
||||||
idx = ogg_find_stream (ogg, serial);
|
idx = ogg_find_stream (ogg, serial);
|
||||||
|
@ -39,7 +39,6 @@ ogm_header(AVFormatContext *s, int idx)
|
|||||||
const uint8_t *p = os->buf + os->pstart;
|
const uint8_t *p = os->buf + os->pstart;
|
||||||
uint64_t time_unit;
|
uint64_t time_unit;
|
||||||
uint64_t spu;
|
uint64_t spu;
|
||||||
uint32_t default_len;
|
|
||||||
|
|
||||||
if(!(*p & 1))
|
if(!(*p & 1))
|
||||||
return 0;
|
return 0;
|
||||||
@ -74,8 +73,7 @@ ogm_header(AVFormatContext *s, int idx)
|
|||||||
|
|
||||||
time_unit = bytestream_get_le64(&p);
|
time_unit = bytestream_get_le64(&p);
|
||||||
spu = bytestream_get_le64(&p);
|
spu = bytestream_get_le64(&p);
|
||||||
default_len = bytestream_get_le32(&p);
|
p += 4; /* default_len */
|
||||||
|
|
||||||
p += 8; /* buffersize + bits_per_sample */
|
p += 8; /* buffersize + bits_per_sample */
|
||||||
|
|
||||||
if(st->codec->codec_type == AVMEDIA_TYPE_VIDEO){
|
if(st->codec->codec_type == AVMEDIA_TYPE_VIDEO){
|
||||||
|
@ -80,8 +80,6 @@ static av_cold int rl2_read_header(AVFormatContext *s,
|
|||||||
unsigned int audio_frame_counter = 0;
|
unsigned int audio_frame_counter = 0;
|
||||||
unsigned int video_frame_counter = 0;
|
unsigned int video_frame_counter = 0;
|
||||||
unsigned int back_size;
|
unsigned int back_size;
|
||||||
int data_size;
|
|
||||||
unsigned short encoding_method;
|
|
||||||
unsigned short sound_rate;
|
unsigned short sound_rate;
|
||||||
unsigned short rate;
|
unsigned short rate;
|
||||||
unsigned short channels;
|
unsigned short channels;
|
||||||
@ -98,14 +96,14 @@ static av_cold int rl2_read_header(AVFormatContext *s,
|
|||||||
avio_skip(pb,4); /* skip FORM tag */
|
avio_skip(pb,4); /* skip FORM tag */
|
||||||
back_size = avio_rl32(pb); /**< get size of the background frame */
|
back_size = avio_rl32(pb); /**< get size of the background frame */
|
||||||
signature = avio_rb32(pb);
|
signature = avio_rb32(pb);
|
||||||
data_size = avio_rb32(pb);
|
avio_skip(pb, 4); /* data size */
|
||||||
frame_count = avio_rl32(pb);
|
frame_count = avio_rl32(pb);
|
||||||
|
|
||||||
/* disallow back_sizes and frame_counts that may lead to overflows later */
|
/* disallow back_sizes and frame_counts that may lead to overflows later */
|
||||||
if(back_size > INT_MAX/2 || frame_count > INT_MAX / sizeof(uint32_t))
|
if(back_size > INT_MAX/2 || frame_count > INT_MAX / sizeof(uint32_t))
|
||||||
return AVERROR_INVALIDDATA;
|
return AVERROR_INVALIDDATA;
|
||||||
|
|
||||||
encoding_method = avio_rl16(pb);
|
avio_skip(pb, 2); /* encoding mentod */
|
||||||
sound_rate = avio_rl16(pb);
|
sound_rate = avio_rl16(pb);
|
||||||
rate = avio_rl16(pb);
|
rate = avio_rl16(pb);
|
||||||
channels = avio_rl16(pb);
|
channels = avio_rl16(pb);
|
||||||
|
@ -280,7 +280,7 @@ ff_rm_read_mdpr_codecdata (AVFormatContext *s, AVIOContext *pb,
|
|||||||
if (rm_read_audio_stream_info(s, pb, st, rst, 0))
|
if (rm_read_audio_stream_info(s, pb, st, rst, 0))
|
||||||
return -1;
|
return -1;
|
||||||
} else {
|
} else {
|
||||||
int fps, fps2;
|
int fps;
|
||||||
if (avio_rl32(pb) != MKTAG('V', 'I', 'D', 'O')) {
|
if (avio_rl32(pb) != MKTAG('V', 'I', 'D', 'O')) {
|
||||||
fail1:
|
fail1:
|
||||||
av_log(st->codec, AV_LOG_ERROR, "Unsupported video codec\n");
|
av_log(st->codec, AV_LOG_ERROR, "Unsupported video codec\n");
|
||||||
@ -298,7 +298,7 @@ ff_rm_read_mdpr_codecdata (AVFormatContext *s, AVIOContext *pb,
|
|||||||
fps= avio_rb16(pb);
|
fps= avio_rb16(pb);
|
||||||
st->codec->codec_type = AVMEDIA_TYPE_VIDEO;
|
st->codec->codec_type = AVMEDIA_TYPE_VIDEO;
|
||||||
avio_rb32(pb);
|
avio_rb32(pb);
|
||||||
fps2= avio_rb16(pb);
|
avio_skip(pb, 2);
|
||||||
avio_rb16(pb);
|
avio_rb16(pb);
|
||||||
|
|
||||||
if ((ret = rm_read_extradata(pb, st->codec, codec_data_size - (avio_tell(pb) - codec_pos))) < 0)
|
if ((ret = rm_read_extradata(pb, st->codec, codec_data_size - (avio_tell(pb) - codec_pos))) < 0)
|
||||||
|
@ -299,9 +299,9 @@ static int rpl_read_packet(AVFormatContext *s, AVPacket *pkt)
|
|||||||
stream->codec->codec_tag == 124) {
|
stream->codec->codec_tag == 124) {
|
||||||
// We have to split Escape 124 frames because there are
|
// We have to split Escape 124 frames because there are
|
||||||
// multiple frames per chunk in Escape 124 samples.
|
// multiple frames per chunk in Escape 124 samples.
|
||||||
uint32_t frame_size, frame_flags;
|
uint32_t frame_size;
|
||||||
|
|
||||||
frame_flags = avio_rl32(pb);
|
avio_skip(pb, 4); /* flags */
|
||||||
frame_size = avio_rl32(pb);
|
frame_size = avio_rl32(pb);
|
||||||
if (avio_seek(pb, -8, SEEK_CUR) < 0)
|
if (avio_seek(pb, -8, SEEK_CUR) < 0)
|
||||||
return AVERROR(EIO);
|
return AVERROR(EIO);
|
||||||
|
@ -108,8 +108,7 @@ static int parse_fmtp_config(AVStream *st, char *value)
|
|||||||
int len = ff_hex_to_data(NULL, value), i, ret = 0;
|
int len = ff_hex_to_data(NULL, value), i, ret = 0;
|
||||||
GetBitContext gb;
|
GetBitContext gb;
|
||||||
uint8_t *config;
|
uint8_t *config;
|
||||||
int audio_mux_version, same_time_framing, num_sub_frames,
|
int audio_mux_version, same_time_framing, num_programs, num_layers;
|
||||||
num_programs, num_layers;
|
|
||||||
|
|
||||||
/* Pad this buffer, too, to avoid out of bounds reads with get_bits below */
|
/* Pad this buffer, too, to avoid out of bounds reads with get_bits below */
|
||||||
config = av_mallocz(len + FF_INPUT_BUFFER_PADDING_SIZE);
|
config = av_mallocz(len + FF_INPUT_BUFFER_PADDING_SIZE);
|
||||||
@ -119,7 +118,7 @@ static int parse_fmtp_config(AVStream *st, char *value)
|
|||||||
init_get_bits(&gb, config, len*8);
|
init_get_bits(&gb, config, len*8);
|
||||||
audio_mux_version = get_bits(&gb, 1);
|
audio_mux_version = get_bits(&gb, 1);
|
||||||
same_time_framing = get_bits(&gb, 1);
|
same_time_framing = get_bits(&gb, 1);
|
||||||
num_sub_frames = get_bits(&gb, 6);
|
skip_bits(&gb, 6); /* num_sub_frames */
|
||||||
num_programs = get_bits(&gb, 4);
|
num_programs = get_bits(&gb, 4);
|
||||||
num_layers = get_bits(&gb, 3);
|
num_layers = get_bits(&gb, 3);
|
||||||
if (audio_mux_version != 0 || same_time_framing != 1 || num_programs != 0 ||
|
if (audio_mux_version != 0 || same_time_framing != 1 || num_programs != 0 ||
|
||||||
|
@ -32,7 +32,7 @@ int ff_sauce_read(AVFormatContext *avctx, uint64_t *fsize, int *got_width, int g
|
|||||||
{
|
{
|
||||||
AVIOContext *pb = avctx->pb;
|
AVIOContext *pb = avctx->pb;
|
||||||
char buf[36];
|
char buf[36];
|
||||||
int datatype, filetype, t1, t2, nb_comments, flags;
|
int datatype, filetype, t1, t2, nb_comments;
|
||||||
uint64_t start_pos = avio_size(pb) - 128;
|
uint64_t start_pos = avio_size(pb) - 128;
|
||||||
|
|
||||||
avio_seek(pb, start_pos, SEEK_SET);
|
avio_seek(pb, start_pos, SEEK_SET);
|
||||||
@ -57,7 +57,7 @@ int ff_sauce_read(AVFormatContext *avctx, uint64_t *fsize, int *got_width, int g
|
|||||||
t1 = avio_rl16(pb);
|
t1 = avio_rl16(pb);
|
||||||
t2 = avio_rl16(pb);
|
t2 = avio_rl16(pb);
|
||||||
nb_comments = avio_r8(pb);
|
nb_comments = avio_r8(pb);
|
||||||
flags = avio_r8(pb);
|
avio_skip(pb, 1); /* flags */
|
||||||
avio_skip(pb, 4);
|
avio_skip(pb, 4);
|
||||||
GET_SAUCE_META("encoder", 22);
|
GET_SAUCE_META("encoder", 22);
|
||||||
|
|
||||||
|
@ -233,7 +233,6 @@ static int smacker_read_packet(AVFormatContext *s, AVPacket *pkt)
|
|||||||
int i;
|
int i;
|
||||||
int frame_size = 0;
|
int frame_size = 0;
|
||||||
int palchange = 0;
|
int palchange = 0;
|
||||||
int pos;
|
|
||||||
|
|
||||||
if (s->pb->eof_reached || smk->cur_frame >= smk->frames)
|
if (s->pb->eof_reached || smk->cur_frame >= smk->frames)
|
||||||
return AVERROR_EOF;
|
return AVERROR_EOF;
|
||||||
@ -244,7 +243,6 @@ static int smacker_read_packet(AVFormatContext *s, AVPacket *pkt)
|
|||||||
frame_size = smk->frm_size[smk->cur_frame] & (~3);
|
frame_size = smk->frm_size[smk->cur_frame] & (~3);
|
||||||
flags = smk->frm_flags[smk->cur_frame];
|
flags = smk->frm_flags[smk->cur_frame];
|
||||||
/* handle palette change event */
|
/* handle palette change event */
|
||||||
pos = avio_tell(s->pb);
|
|
||||||
if(flags & SMACKER_PAL){
|
if(flags & SMACKER_PAL){
|
||||||
int size, sz, t, off, j, pos;
|
int size, sz, t, off, j, pos;
|
||||||
uint8_t *pal = smk->pal;
|
uint8_t *pal = smk->pal;
|
||||||
|
@ -85,7 +85,6 @@ static int sol_channels(int magic, int type)
|
|||||||
static int sol_read_header(AVFormatContext *s,
|
static int sol_read_header(AVFormatContext *s,
|
||||||
AVFormatParameters *ap)
|
AVFormatParameters *ap)
|
||||||
{
|
{
|
||||||
int size;
|
|
||||||
unsigned int magic,tag;
|
unsigned int magic,tag;
|
||||||
AVIOContext *pb = s->pb;
|
AVIOContext *pb = s->pb;
|
||||||
unsigned int id, channels, rate, type;
|
unsigned int id, channels, rate, type;
|
||||||
@ -99,7 +98,7 @@ static int sol_read_header(AVFormatContext *s,
|
|||||||
return -1;
|
return -1;
|
||||||
rate = avio_rl16(pb);
|
rate = avio_rl16(pb);
|
||||||
type = avio_r8(pb);
|
type = avio_r8(pb);
|
||||||
size = avio_rl32(pb);
|
avio_skip(pb, 4); /* size */
|
||||||
if (magic != 0x0B8D)
|
if (magic != 0x0B8D)
|
||||||
avio_r8(pb); /* newer SOLs contain padding byte */
|
avio_r8(pb); /* newer SOLs contain padding byte */
|
||||||
|
|
||||||
|
@ -94,7 +94,7 @@ static int yuv4_write_packet(AVFormatContext *s, AVPacket *pkt)
|
|||||||
AVPicture *picture;
|
AVPicture *picture;
|
||||||
int* first_pkt = s->priv_data;
|
int* first_pkt = s->priv_data;
|
||||||
int width, height, h_chroma_shift, v_chroma_shift;
|
int width, height, h_chroma_shift, v_chroma_shift;
|
||||||
int i, m;
|
int i;
|
||||||
char buf2[Y4M_LINE_MAX+1];
|
char buf2[Y4M_LINE_MAX+1];
|
||||||
char buf1[20];
|
char buf1[20];
|
||||||
uint8_t *ptr, *ptr1, *ptr2;
|
uint8_t *ptr, *ptr1, *ptr2;
|
||||||
@ -114,7 +114,7 @@ static int yuv4_write_packet(AVFormatContext *s, AVPacket *pkt)
|
|||||||
|
|
||||||
/* construct frame header */
|
/* construct frame header */
|
||||||
|
|
||||||
m = snprintf(buf1, sizeof(buf1), "%s\n", Y4M_FRAME_MAGIC);
|
snprintf(buf1, sizeof(buf1), "%s\n", Y4M_FRAME_MAGIC);
|
||||||
avio_write(pb, buf1, strlen(buf1));
|
avio_write(pb, buf1, strlen(buf1));
|
||||||
|
|
||||||
width = st->codec->width;
|
width = st->codec->width;
|
||||||
|
Loading…
Reference in New Issue
Block a user