1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00

avformat: Avoid allocation for AVStreamInternal

Do this by allocating AVStream together with the data that is
currently in AVStreamInternal; or rather: Put AVStream at the
beginning of a new structure called FFStream (which encompasses
more than just the internal fields and is a proper context in its own
right, hence the name) and remove AVStreamInternal altogether.

Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
This commit is contained in:
Andreas Rheinhardt 2021-08-24 19:41:16 +02:00
parent 9f05b3ba60
commit 40bdd8cc05
87 changed files with 745 additions and 606 deletions

View File

@ -112,7 +112,7 @@ static int adts_aac_read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = s->iformat->raw_codec_id;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ff_id3v1_read(s);
if ((s->pb->seekable & AVIO_SEEKABLE_NORMAL) &&

View File

@ -86,6 +86,7 @@ static int aa_read_header(AVFormatContext *s)
AADemuxContext *c = s->priv_data;
AVIOContext *pb = s->pb;
AVStream *st;
FFStream *sti;
int ret;
/* parse .aa header */
@ -178,11 +179,12 @@ static int aa_read_header(AVFormatContext *s)
st = avformat_new_stream(s, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
if (!strcmp(codec_name, "mp332")) {
st->codecpar->codec_id = AV_CODEC_ID_MP3;
st->codecpar->sample_rate = 22050;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 8, 32000 * TIMEPREC);
// encoded audio frame is MP3_FRAME_SIZE bytes (+1 with padding, unlikely)
} else if (!strcmp(codec_name, "acelp85")) {
@ -191,7 +193,7 @@ static int aa_read_header(AVFormatContext *s)
st->codecpar->channels = 1;
st->codecpar->sample_rate = 8500;
st->codecpar->bit_rate = 8500;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 8, 8500 * TIMEPREC);
} else if (!strcmp(codec_name, "acelp16")) {
st->codecpar->codec_id = AV_CODEC_ID_SIPR;
@ -199,7 +201,7 @@ static int aa_read_header(AVFormatContext *s)
st->codecpar->channels = 1;
st->codecpar->sample_rate = 16000;
st->codecpar->bit_rate = 16000;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 8, 16000 * TIMEPREC);
}

View File

@ -54,7 +54,7 @@ static int acm_read_header(AVFormatContext *s)
return AVERROR_INVALIDDATA;
st->start_time = 0;
st->duration = AV_RL32(st->codecpar->extradata + 4) / st->codecpar->channels;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 1, st->codecpar->sample_rate);
return 0;

View File

@ -434,7 +434,7 @@ static int ape_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
if (index < 0)
return -1;
if ((ret = avio_seek(s->pb, st->internal->index_entries[index].pos, SEEK_SET)) < 0)
if ((ret = avio_seek(s->pb, ffstream(st)->index_entries[index].pos, SEEK_SET)) < 0)
return ret;
ape->currentframe = index;
return 0;

View File

@ -299,6 +299,7 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
ASFContext *asf = s->priv_data;
AVIOContext *pb = s->pb;
AVStream *st;
FFStream *sti;
ASFStream *asf_st;
ff_asf_guid g;
enum AVMediaType type;
@ -317,6 +318,7 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
st = avformat_new_stream(s, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
avpriv_set_pts_info(st, 32, 1, 1000); /* 32 bit pts in ms */
start_time = asf->hdr.preroll;
@ -378,13 +380,13 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
if (is_dvr_ms_audio) {
// codec_id and codec_tag are unreliable in dvr_ms
// files. Set them later by probing stream.
st->internal->request_probe = 1;
sti->request_probe = 1;
st->codecpar->codec_tag = 0;
}
if (st->codecpar->codec_id == AV_CODEC_ID_AAC)
st->internal->need_parsing = AVSTREAM_PARSE_NONE;
sti->need_parsing = AVSTREAM_PARSE_NONE;
else
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
/* We have to init the frame size at some point .... */
pos2 = avio_tell(pb);
if (size >= (pos2 + 8 - pos1 + 24)) {
@ -443,7 +445,7 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
st->codecpar->codec_tag = tag1;
st->codecpar->codec_id = ff_codec_get_id(ff_codec_bmp_tags, tag1);
if (tag1 == MKTAG('D', 'V', 'R', ' ')) {
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
/* issue658 contains wrong w/h and MS even puts a fake seq header
* with wrong w/h in extradata while a correct one is in the stream.
* maximum lameness */
@ -453,9 +455,9 @@ static int asf_read_stream_properties(AVFormatContext *s, int64_t size)
st->codecpar->extradata_size = 0;
}
if (st->codecpar->codec_id == AV_CODEC_ID_H264)
st->internal->need_parsing = AVSTREAM_PARSE_FULL_ONCE;
sti->need_parsing = AVSTREAM_PARSE_FULL_ONCE;
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG4)
st->internal->need_parsing = AVSTREAM_PARSE_FULL_ONCE;
sti->need_parsing = AVSTREAM_PARSE_FULL_ONCE;
}
pos2 = avio_tell(pb);
avio_skip(pb, size - (pos2 - pos1 + 24));
@ -1557,6 +1559,7 @@ static int asf_read_seek(AVFormatContext *s, int stream_index,
{
ASFContext *asf = s->priv_data;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int ret = 0;
if (s->packet_size <= 0)
@ -1584,11 +1587,11 @@ static int asf_read_seek(AVFormatContext *s, int stream_index,
asf->index_read = -1;
}
if (asf->index_read > 0 && st->internal->index_entries) {
if (asf->index_read > 0 && sti->index_entries) {
int index = av_index_search_timestamp(st, pts, flags);
if (index >= 0) {
/* find the position */
uint64_t pos = st->internal->index_entries[index].pos;
uint64_t pos = sti->index_entries[index].pos;
/* do the seek */
av_log(s, AV_LOG_DEBUG, "SEEKTO: %"PRId64"\n", pos);

View File

@ -1513,13 +1513,15 @@ static int asf_read_seek(AVFormatContext *s, int stream_index,
int64_t timestamp, int flags)
{
ASFContext *asf = s->priv_data;
AVStream *const st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int idx, ret;
if (s->streams[stream_index]->internal->nb_index_entries && asf->is_simple_index) {
idx = av_index_search_timestamp(s->streams[stream_index], timestamp, flags);
if (idx < 0 || idx >= s->streams[stream_index]->internal->nb_index_entries)
if (sti->nb_index_entries && asf->is_simple_index) {
idx = av_index_search_timestamp(st, timestamp, flags);
if (idx < 0 || idx >= sti->nb_index_entries)
return AVERROR_INVALIDDATA;
avio_seek(s->pb, s->streams[stream_index]->internal->index_entries[idx].pos, SEEK_SET);
avio_seek(s->pb, sti->index_entries[idx].pos, SEEK_SET);
} else {
if ((ret = ff_seek_frame_binary(s, stream_index, timestamp, flags)) < 0)
return ret;

View File

@ -61,6 +61,7 @@ static int av1_read_header(AVFormatContext *s)
AV1DemuxContext *const c = s->priv_data;
const AVBitStreamFilter *filter = av_bsf_get_by_name("av1_frame_merge");
AVStream *st;
FFStream *sti;
int ret;
if (!filter) {
@ -72,12 +73,13 @@ static int av1_read_header(AVFormatContext *s)
st = avformat_new_stream(s, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_AV1;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
st->internal->avctx->framerate = c->framerate;
sti->avctx->framerate = c->framerate;
// taken from rawvideo demuxers
avpriv_set_pts_info(st, 64, 1, 1200000);

View File

@ -821,8 +821,6 @@ typedef struct AVIndexEntry {
*/
#define AV_DISPOSITION_TIMED_THUMBNAILS 0x0800
typedef struct AVStreamInternal AVStreamInternal;
/**
* To specify text track kind (different from subtitles default).
*/
@ -1003,12 +1001,6 @@ typedef struct AVStream {
*
*/
int pts_wrap_bits;
/**
* An opaque field for libavformat internal usage.
* Must not be accessed in any way by callers.
*/
AVStreamInternal *internal;
} AVStream;
struct AVCodecParserContext *av_stream_get_parser(const AVStream *s);

View File

@ -279,8 +279,9 @@ static void clean_index(AVFormatContext *s)
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
AVIStream *ast = st->priv_data;
int n = st->internal->nb_index_entries;
int n = sti->nb_index_entries;
int max = ast->sample_size;
int64_t pos, size, ts;
@ -290,9 +291,9 @@ static void clean_index(AVFormatContext *s)
while (max < 1024)
max += max;
pos = st->internal->index_entries[0].pos;
size = st->internal->index_entries[0].size;
ts = st->internal->index_entries[0].timestamp;
pos = sti->index_entries[0].pos;
size = sti->index_entries[0].size;
ts = sti->index_entries[0].timestamp;
for (j = 0; j < size; j += max)
av_add_index_entry(st, pos + j, ts + j, FFMIN(max, size - j), 0,
@ -439,14 +440,14 @@ static int calculate_bitrate(AVFormatContext *s)
for (i = 0; i<s->nb_streams; i++) {
int64_t len = 0;
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(s->streams[i]);
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
continue;
for (j = 0; j < st->internal->nb_index_entries; j++)
len += st->internal->index_entries[j].size;
maxpos = FFMAX(maxpos, st->internal->index_entries[j-1].pos);
for (j = 0; j < sti->nb_index_entries; j++)
len += sti->index_entries[j].size;
maxpos = FFMAX(maxpos, sti->index_entries[j-1].pos);
lensum += len;
}
if (maxpos < av_rescale(avi->io_fsize, 9, 10)) // index does not cover the whole file
@ -457,15 +458,16 @@ static int calculate_bitrate(AVFormatContext *s)
for (i = 0; i<s->nb_streams; i++) {
int64_t len = 0;
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
int64_t duration;
int64_t bitrate;
for (j = 0; j < st->internal->nb_index_entries; j++)
len += st->internal->index_entries[j].size;
for (j = 0; j < sti->nb_index_entries; j++)
len += sti->index_entries[j].size;
if (st->internal->nb_index_entries < 2 || st->codecpar->bit_rate > 0)
if (sti->nb_index_entries < 2 || st->codecpar->bit_rate > 0)
continue;
duration = st->internal->index_entries[j-1].timestamp - st->internal->index_entries[0].timestamp;
duration = sti->index_entries[j-1].timestamp - sti->index_entries[0].timestamp;
bitrate = av_rescale(8*len, st->time_base.den, duration * st->time_base.num);
if (bitrate > 0) {
st->codecpar->bit_rate = bitrate;
@ -736,10 +738,12 @@ static int avi_read_header(AVFormatContext *s)
avio_skip(pb, size);
} else {
uint64_t cur_pos = avio_tell(pb);
FFStream *sti;
unsigned esize;
if (cur_pos < list_end)
size = FFMIN(size, list_end - cur_pos);
st = s->streams[stream_index];
sti = ffstream(st);
if (st->codecpar->codec_type != AVMEDIA_TYPE_UNKNOWN) {
avio_skip(pb, size);
break;
@ -823,19 +827,19 @@ static int avi_read_header(AVFormatContext *s)
/* This is needed to get the pict type which is necessary
* for generating correct pts. */
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG4 &&
ast->handler == MKTAG('X', 'V', 'I', 'D'))
st->codecpar->codec_tag = MKTAG('X', 'V', 'I', 'D');
if (st->codecpar->codec_tag == MKTAG('V', 'S', 'S', 'H'))
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
if (st->codecpar->codec_id == AV_CODEC_ID_RV40)
st->internal->need_parsing = AVSTREAM_PARSE_NONE;
sti->need_parsing = AVSTREAM_PARSE_NONE;
if (st->codecpar->codec_id == AV_CODEC_ID_HEVC &&
st->codecpar->codec_tag == MKTAG('H', '2', '6', '5'))
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
if (st->codecpar->codec_id == AV_CODEC_ID_AVRN &&
st->codecpar->codec_tag == MKTAG('A', 'V', 'R', 'n') &&
@ -879,16 +883,16 @@ static int avi_read_header(AVFormatContext *s)
avio_skip(pb, 1);
/* Force parsing as several audio frames can be in
* one packet and timestamps refer to packet start. */
st->internal->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
sti->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
/* ADTS header is in extradata, AAC without header must be
* stored as exact frames. Parser not needed and it will
* fail. */
if (st->codecpar->codec_id == AV_CODEC_ID_AAC &&
st->codecpar->extradata_size)
st->internal->need_parsing = AVSTREAM_PARSE_NONE;
sti->need_parsing = AVSTREAM_PARSE_NONE;
// The flac parser does not work with AVSTREAM_PARSE_TIMESTAMPS
if (st->codecpar->codec_id == AV_CODEC_ID_FLAC)
st->internal->need_parsing = AVSTREAM_PARSE_NONE;
sti->need_parsing = AVSTREAM_PARSE_NONE;
/* AVI files with Xan DPCM audio (wrongly) declare PCM
* audio in the header but have Axan as stream_code_tag. */
if (ast->handler == AV_RL32("Axan")) {
@ -915,7 +919,7 @@ static int avi_read_header(AVFormatContext *s)
break;
case AVMEDIA_TYPE_SUBTITLE:
st->codecpar->codec_type = AVMEDIA_TYPE_SUBTITLE;
st->internal->request_probe= 1;
sti->request_probe = 1;
avio_skip(pb, size);
break;
default:
@ -1050,12 +1054,12 @@ end_of_header:
AVStream *st = s->streams[i];
if ( st->codecpar->codec_id == AV_CODEC_ID_MPEG1VIDEO
|| st->codecpar->codec_id == AV_CODEC_ID_MPEG2VIDEO)
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
}
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
if (st->internal->nb_index_entries)
if (ffstream(st)->nb_index_entries)
break;
}
// DV-in-AVI cannot be non-interleaved, if set this must be
@ -1341,9 +1345,10 @@ start_sync:
ast->remaining = size;
if (size) {
FFStream *const sti = ffstream(st);
uint64_t pos = avio_tell(pb) - 8;
if (!st->internal->index_entries || !st->internal->nb_index_entries ||
st->internal->index_entries[st->internal->nb_index_entries - 1].pos < pos) {
if (!sti->index_entries || !sti->nb_index_entries ||
sti->index_entries[sti->nb_index_entries - 1].pos < pos) {
av_add_index_entry(st, pos, ast->frame_offset, size,
0, AVINDEX_KEYFRAME);
}
@ -1363,20 +1368,22 @@ static int ni_prepare_read(AVFormatContext *s)
AVIContext *avi = s->priv_data;
int best_stream_index = 0;
AVStream *best_st = NULL;
FFStream *best_sti;
AVIStream *best_ast;
int64_t best_ts = INT64_MAX;
int i;
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
AVIStream *ast = st->priv_data;
int64_t ts = ast->frame_offset;
int64_t last_ts;
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
continue;
last_ts = st->internal->index_entries[st->internal->nb_index_entries - 1].timestamp;
last_ts = sti->index_entries[sti->nb_index_entries - 1].timestamp;
if (!ast->remaining && ts > last_ts)
continue;
@ -1395,6 +1402,7 @@ static int ni_prepare_read(AVFormatContext *s)
if (!best_st)
return AVERROR_EOF;
best_sti = ffstream(best_st);
best_ast = best_st->priv_data;
best_ts = best_ast->frame_offset;
if (best_ast->remaining) {
@ -1405,11 +1413,11 @@ static int ni_prepare_read(AVFormatContext *s)
} else {
i = av_index_search_timestamp(best_st, best_ts, AVSEEK_FLAG_ANY);
if (i >= 0)
best_ast->frame_offset = best_st->internal->index_entries[i].timestamp;
best_ast->frame_offset = best_sti->index_entries[i].timestamp;
}
if (i >= 0) {
int64_t pos = best_st->internal->index_entries[i].pos;
int64_t pos = best_sti->index_entries[i].pos;
pos += best_ast->packet_size - best_ast->remaining;
if (avio_seek(s->pb, pos + 8, SEEK_SET) < 0)
return AVERROR_EOF;
@ -1419,7 +1427,7 @@ static int ni_prepare_read(AVFormatContext *s)
avi->stream_index = best_stream_index;
if (!best_ast->remaining)
best_ast->packet_size =
best_ast->remaining = best_st->internal->index_entries[i].size;
best_ast->remaining = best_sti->index_entries[i].size;
}
else
return AVERROR_EOF;
@ -1450,6 +1458,7 @@ static int avi_read_packet(AVFormatContext *s, AVPacket *pkt)
resync:
if (avi->stream_index >= 0) {
AVStream *st = s->streams[avi->stream_index];
FFStream *const sti = ffstream(st);
AVIStream *ast = st->priv_data;
int dv_demux = CONFIG_DV_DEMUXER && avi->dv_demux;
int size, err;
@ -1508,15 +1517,15 @@ resync:
pkt->dts /= ast->sample_size;
pkt->stream_index = avi->stream_index;
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && st->internal->index_entries) {
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && sti->index_entries) {
AVIndexEntry *e;
int index;
index = av_index_search_timestamp(st, ast->frame_offset, AVSEEK_FLAG_ANY);
e = &st->internal->index_entries[index];
e = &sti->index_entries[index];
if (index >= 0 && e->timestamp == ast->frame_offset) {
if (index == st->internal->nb_index_entries-1) {
if (index == sti->nb_index_entries-1) {
int key=1;
uint32_t state=-1;
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG4) {
@ -1552,7 +1561,7 @@ resync:
}
ast->seek_pos= 0;
if (!avi->non_interleaved && st->internal->nb_index_entries>1 && avi->index_loaded>1) {
if (!avi->non_interleaved && sti->nb_index_entries > 1 && avi->index_loaded > 1) {
int64_t dts= av_rescale_q(pkt->dts, st->time_base, AV_TIME_BASE_Q);
if (avi->dts_max < dts) {
@ -1651,9 +1660,9 @@ static int avi_read_idx1(AVFormatContext *s, int size)
}
if (!anykey) {
for (index = 0; index < s->nb_streams; index++) {
st = s->streams[index];
if (st->internal->nb_index_entries)
st->internal->index_entries[0].flags |= AVINDEX_KEYFRAME;
FFStream *const sti = ffstream(s->streams[index]);
if (sti->nb_index_entries)
sti->index_entries[0].flags |= AVINDEX_KEYFRAME;
}
}
return 0;
@ -1678,25 +1687,27 @@ static int check_stream_max_drift(AVFormatContext *s)
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
AVIStream *ast = st->priv_data;
int n = st->internal->nb_index_entries;
while (idx[i] < n && st->internal->index_entries[idx[i]].pos < pos)
FFStream *const sti = ffstream(st);
int n = sti->nb_index_entries;
while (idx[i] < n && sti->index_entries[idx[i]].pos < pos)
idx[i]++;
if (idx[i] < n) {
int64_t dts;
dts = av_rescale_q(st->internal->index_entries[idx[i]].timestamp /
dts = av_rescale_q(sti->index_entries[idx[i]].timestamp /
FFMAX(ast->sample_size, 1),
st->time_base, AV_TIME_BASE_Q);
min_dts = FFMIN(min_dts, dts);
min_pos = FFMIN(min_pos, st->internal->index_entries[idx[i]].pos);
min_pos = FFMIN(min_pos, sti->index_entries[idx[i]].pos);
}
}
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
AVIStream *ast = st->priv_data;
if (idx[i] && min_dts != INT64_MAX / 2) {
int64_t dts, delta_dts;
dts = av_rescale_q(st->internal->index_entries[idx[i] - 1].timestamp /
dts = av_rescale_q(sti->index_entries[idx[i] - 1].timestamp /
FFMAX(ast->sample_size, 1),
st->time_base, AV_TIME_BASE_Q);
delta_dts = av_sat_sub64(dts, min_dts);
@ -1726,30 +1737,31 @@ static int guess_ni_flag(AVFormatContext *s)
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
int n = st->internal->nb_index_entries;
FFStream *const sti = ffstream(st);
int n = sti->nb_index_entries;
unsigned int size;
if (n <= 0)
continue;
if (n >= 2) {
int64_t pos = st->internal->index_entries[0].pos;
int64_t pos = sti->index_entries[0].pos;
unsigned tag[2];
avio_seek(s->pb, pos, SEEK_SET);
tag[0] = avio_r8(s->pb);
tag[1] = avio_r8(s->pb);
avio_rl16(s->pb);
size = avio_rl32(s->pb);
if (get_stream_idx(tag) == i && pos + size > st->internal->index_entries[1].pos)
if (get_stream_idx(tag) == i && pos + size > sti->index_entries[1].pos)
last_start = INT64_MAX;
if (get_stream_idx(tag) == i && size == st->internal->index_entries[0].size + 8)
if (get_stream_idx(tag) == i && size == sti->index_entries[0].size + 8)
last_start = INT64_MAX;
}
if (st->internal->index_entries[0].pos > last_start)
last_start = st->internal->index_entries[0].pos;
if (st->internal->index_entries[n - 1].pos < first_end)
first_end = st->internal->index_entries[n - 1].pos;
if (sti->index_entries[0].pos > last_start)
last_start = sti->index_entries[0].pos;
if (sti->index_entries[n - 1].pos < first_end)
first_end = sti->index_entries[n - 1].pos;
}
avio_seek(s->pb, oldpos, SEEK_SET);
@ -1817,6 +1829,7 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
{
AVIContext *avi = s->priv_data;
AVStream *st;
FFStream *sti;
int i, index;
int64_t pos, pos_min;
AVIStream *ast;
@ -1835,25 +1848,26 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
av_assert0(stream_index >= 0);
st = s->streams[stream_index];
sti = ffstream(st);
ast = st->priv_data;
index = av_index_search_timestamp(st,
timestamp * FFMAX(ast->sample_size, 1),
flags);
if (index < 0) {
if (st->internal->nb_index_entries > 0)
if (sti->nb_index_entries > 0)
av_log(s, AV_LOG_DEBUG, "Failed to find timestamp %"PRId64 " in index %"PRId64 " .. %"PRId64 "\n",
timestamp * FFMAX(ast->sample_size, 1),
st->internal->index_entries[0].timestamp,
st->internal->index_entries[st->internal->nb_index_entries - 1].timestamp);
sti->index_entries[0].timestamp,
sti->index_entries[sti->nb_index_entries - 1].timestamp);
return AVERROR_INVALIDDATA;
}
/* find the position */
pos = st->internal->index_entries[index].pos;
timestamp = st->internal->index_entries[index].timestamp / FFMAX(ast->sample_size, 1);
pos = sti->index_entries[index].pos;
timestamp = sti->index_entries[index].timestamp / FFMAX(ast->sample_size, 1);
av_log(s, AV_LOG_TRACE, "XX %"PRId64" %d %"PRId64"\n",
timestamp, index, st->internal->index_entries[index].timestamp);
timestamp, index, sti->index_entries[index].timestamp);
if (CONFIG_DV_DEMUXER && avi->dv_demux) {
/* One and only one real stream for DV in AVI, and it has video */
@ -1874,6 +1888,7 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
pos_min = pos;
for (i = 0; i < s->nb_streams; i++) {
AVStream *st2 = s->streams[i];
FFStream *const sti2 = ffstream(st2);
AVIStream *ast2 = st2->priv_data;
ast2->packet_size =
@ -1884,7 +1899,7 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
continue;
}
if (st2->internal->nb_index_entries <= 0)
if (sti2->nb_index_entries <= 0)
continue;
// av_assert1(st2->codecpar->block_align);
@ -1898,14 +1913,15 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
(st2->codecpar->codec_type != AVMEDIA_TYPE_VIDEO ? AVSEEK_FLAG_ANY : 0));
if (index < 0)
index = 0;
ast2->seek_pos = st2->internal->index_entries[index].pos;
ast2->seek_pos = sti2->index_entries[index].pos;
pos_min = FFMIN(pos_min,ast2->seek_pos);
}
for (i = 0; i < s->nb_streams; i++) {
AVStream *st2 = s->streams[i];
FFStream *const sti2 = ffstream(st2);
AVIStream *ast2 = st2->priv_data;
if (ast2->sub_ctx || st2->internal->nb_index_entries <= 0)
if (ast2->sub_ctx || sti2->nb_index_entries <= 0)
continue;
index = av_index_search_timestamp(
@ -1914,9 +1930,9 @@ static int avi_read_seek(AVFormatContext *s, int stream_index,
flags | AVSEEK_FLAG_BACKWARD | (st2->codecpar->codec_type != AVMEDIA_TYPE_VIDEO ? AVSEEK_FLAG_ANY : 0));
if (index < 0)
index = 0;
while (!avi->non_interleaved && index>0 && st2->internal->index_entries[index-1].pos >= pos_min)
while (!avi->non_interleaved && index > 0 && sti2->index_entries[index-1].pos >= pos_min)
index--;
ast2->frame_offset = st2->internal->index_entries[index].timestamp;
ast2->frame_offset = sti2->index_entries[index].timestamp;
}
/* do the seek */

View File

@ -87,7 +87,8 @@ static int read_header(AVFormatContext *s)
BinkDemuxContext *bink = s->priv_data;
AVIOContext *pb = s->pb;
uint32_t fps_num, fps_den;
AVStream *vst, *ast;
AVStream *const vst = avformat_new_stream(s, NULL);
FFStream *const vsti = ffstream(vst);
unsigned int i;
uint32_t pos, next_pos;
uint16_t flags;
@ -97,7 +98,6 @@ static int read_header(AVFormatContext *s)
uint32_t signature;
uint8_t revision;
vst = avformat_new_stream(s, NULL);
if (!vst)
return AVERROR(ENOMEM);
@ -175,7 +175,7 @@ static int read_header(AVFormatContext *s)
avio_skip(pb, 4 * bink->num_audio_tracks); /* max decoded size */
for (i = 0; i < bink->num_audio_tracks; i++) {
ast = avformat_new_stream(s, NULL);
AVStream *const ast = avformat_new_stream(s, NULL);
if (!ast)
return AVERROR(ENOMEM);
ast->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
@ -225,8 +225,8 @@ static int read_header(AVFormatContext *s)
return ret;
}
if (vst->internal->index_entries)
avio_seek(pb, vst->internal->index_entries[0].pos + bink->smush_size, SEEK_SET);
if (vsti->index_entries)
avio_seek(pb, vsti->index_entries[0].pos + bink->smush_size, SEEK_SET);
else
avio_skip(pb, 4);
@ -243,6 +243,7 @@ static int read_packet(AVFormatContext *s, AVPacket *pkt)
if (bink->current_track < 0) {
int index_entry;
AVStream *st = s->streams[0]; // stream 0 is video stream with index
FFStream *const sti = ffstream(st);
if (bink->video_pts >= st->duration)
return AVERROR_EOF;
@ -256,8 +257,8 @@ static int read_packet(AVFormatContext *s, AVPacket *pkt)
return AVERROR(EIO);
}
bink->remain_packet_size = st->internal->index_entries[index_entry].size;
bink->flags = st->internal->index_entries[index_entry].flags;
bink->remain_packet_size = sti->index_entries[index_entry].size;
bink->flags = sti->index_entries[index_entry].flags;
bink->current_track = 0;
}
@ -307,13 +308,14 @@ static int read_seek(AVFormatContext *s, int stream_index, int64_t timestamp, in
{
BinkDemuxContext *bink = s->priv_data;
AVStream *vst = s->streams[0];
FFStream *const vsti = ffstream(vst);
int64_t ret;
if (!(s->pb->seekable & AVIO_SEEKABLE_NORMAL))
return -1;
/* seek to the first frame */
ret = avio_seek(s->pb, vst->internal->index_entries[0].pos + bink->smush_size, SEEK_SET);
ret = avio_seek(s->pb, vsti->index_entries[0].pos + bink->smush_size, SEEK_SET);
if (ret < 0)
return ret;

View File

@ -342,7 +342,7 @@ found_data:
if (caf->bytes_per_packet > 0 && caf->frames_per_packet > 0) {
if (caf->data_size > 0)
st->nb_frames = (caf->data_size / caf->bytes_per_packet) * caf->frames_per_packet;
} else if (st->internal->nb_index_entries && st->duration > 0) {
} else if (ffstream(st)->nb_index_entries && st->duration > 0) {
if (st->codecpar->sample_rate && caf->data_size / st->duration > INT64_MAX / st->codecpar->sample_rate / 8) {
av_log(s, AV_LOG_ERROR, "Overflow during bit rate calculation %d * 8 * %"PRId64"\n",
st->codecpar->sample_rate, caf->data_size / st->duration);
@ -372,6 +372,7 @@ static int read_packet(AVFormatContext *s, AVPacket *pkt)
{
AVIOContext *pb = s->pb;
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
CafContext *caf = s->priv_data;
int res, pkt_size = 0, pkt_frames = 0;
int64_t left = CAF_MAX_PKT_SIZE;
@ -395,13 +396,13 @@ static int read_packet(AVFormatContext *s, AVPacket *pkt)
pkt_size = (CAF_MAX_PKT_SIZE / pkt_size) * pkt_size;
pkt_size = FFMIN(pkt_size, left);
pkt_frames = pkt_size / caf->bytes_per_packet;
} else if (st->internal->nb_index_entries) {
if (caf->packet_cnt < st->internal->nb_index_entries - 1) {
pkt_size = st->internal->index_entries[caf->packet_cnt + 1].pos - st->internal->index_entries[caf->packet_cnt].pos;
pkt_frames = st->internal->index_entries[caf->packet_cnt + 1].timestamp - st->internal->index_entries[caf->packet_cnt].timestamp;
} else if (caf->packet_cnt == st->internal->nb_index_entries - 1) {
pkt_size = caf->num_bytes - st->internal->index_entries[caf->packet_cnt].pos;
pkt_frames = st->duration - st->internal->index_entries[caf->packet_cnt].timestamp;
} else if (sti->nb_index_entries) {
if (caf->packet_cnt < sti->nb_index_entries - 1) {
pkt_size = sti->index_entries[caf->packet_cnt + 1].pos - sti->index_entries[caf->packet_cnt].pos;
pkt_frames = sti->index_entries[caf->packet_cnt + 1].timestamp - sti->index_entries[caf->packet_cnt].timestamp;
} else if (caf->packet_cnt == sti->nb_index_entries - 1) {
pkt_size = caf->num_bytes - sti->index_entries[caf->packet_cnt].pos;
pkt_frames = st->duration - sti->index_entries[caf->packet_cnt].timestamp;
} else {
return AVERROR(EIO);
}
@ -428,6 +429,7 @@ static int read_seek(AVFormatContext *s, int stream_index,
int64_t timestamp, int flags)
{
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
CafContext *caf = s->priv_data;
int64_t pos, packet_cnt, frame_cnt;
@ -440,10 +442,10 @@ static int read_seek(AVFormatContext *s, int stream_index,
pos = FFMIN(pos, caf->data_size);
packet_cnt = pos / caf->bytes_per_packet;
frame_cnt = caf->frames_per_packet * packet_cnt;
} else if (st->internal->nb_index_entries) {
} else if (sti->nb_index_entries) {
packet_cnt = av_index_search_timestamp(st, timestamp, flags);
frame_cnt = st->internal->index_entries[packet_cnt].timestamp;
pos = st->internal->index_entries[packet_cnt].pos;
frame_cnt = sti->index_entries[packet_cnt].timestamp;
pos = sti->index_entries[packet_cnt].pos;
} else {
return -1;
}

View File

@ -285,13 +285,14 @@ static int cine_read_packet(AVFormatContext *avctx, AVPacket *pkt)
{
CineDemuxContext *cine = avctx->priv_data;
AVStream *st = avctx->streams[0];
FFStream *const sti = ffstream(st);
AVIOContext *pb = avctx->pb;
int n, size, ret;
if (cine->pts >= st->internal->nb_index_entries)
if (cine->pts >= sti->nb_index_entries)
return AVERROR_EOF;
avio_seek(pb, st->internal->index_entries[cine->pts].pos, SEEK_SET);
avio_seek(pb, sti->index_entries[cine->pts].pos, SEEK_SET);
n = avio_rl32(pb);
if (n < 8)
return AVERROR_INVALIDDATA;

View File

@ -738,6 +738,7 @@ static int concat_read_packet(AVFormatContext *avf, AVPacket *pkt)
int64_t delta;
ConcatStream *cs;
AVStream *st;
FFStream *sti;
if (cat->eof)
return AVERROR_EOF;
@ -774,6 +775,7 @@ static int concat_read_packet(AVFormatContext *avf, AVPacket *pkt)
return ret;
st = cat->avf->streams[pkt->stream_index];
sti = ffstream(st);
av_log(avf, AV_LOG_DEBUG, "file:%d stream:%d pts:%s pts_time:%s dts:%s dts_time:%s",
(unsigned)(cat->cur_file - cat->files), pkt->stream_index,
av_ts2str(pkt->pts), av_ts2timestr(pkt->pts, &st->time_base),
@ -802,8 +804,8 @@ static int concat_read_packet(AVFormatContext *avf, AVPacket *pkt)
}
}
if (cat->cur_file->duration == AV_NOPTS_VALUE && st->internal->cur_dts != AV_NOPTS_VALUE) {
int64_t next_dts = av_rescale_q(st->internal->cur_dts, st->time_base, AV_TIME_BASE_Q);
if (cat->cur_file->duration == AV_NOPTS_VALUE && sti->cur_dts != AV_NOPTS_VALUE) {
int64_t next_dts = av_rescale_q(sti->cur_dts, st->time_base, AV_TIME_BASE_Q);
if (cat->cur_file->next_dts == AV_NOPTS_VALUE || next_dts > cat->cur_file->next_dts) {
cat->cur_file->next_dts = next_dts;
}

View File

@ -2344,10 +2344,10 @@ static int dash_check_bitstream(struct AVFormatContext *s, const AVPacket *avpkt
pkt.stream_index = 0;
ret = oc->oformat->check_bitstream(oc, &pkt);
if (ret == 1) {
AVStream *st = s->streams[avpkt->stream_index];
AVStream *ost = oc->streams[0];
st->internal->bsfc = ost->internal->bsfc;
ost->internal->bsfc = NULL;
FFStream *const sti = ffstream(s->streams[avpkt->stream_index]);
FFStream *const osti = ffstream(oc->streams[0]);
sti->bsfc = osti->bsfc;
osti->bsfc = NULL;
}
return ret;
}

View File

@ -433,15 +433,16 @@ static int dhav_read_seek(AVFormatContext *s, int stream_index,
{
DHAVContext *dhav = s->priv_data;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int index = av_index_search_timestamp(st, timestamp, flags);
int64_t pts;
if (index < 0)
return -1;
if (avio_seek(s->pb, st->internal->index_entries[index].pos, SEEK_SET) < 0)
if (avio_seek(s->pb, sti->index_entries[index].pos, SEEK_SET) < 0)
return -1;
pts = st->internal->index_entries[index].timestamp;
pts = sti->index_entries[index].timestamp;
for (int n = 0; n < s->nb_streams; n++) {
AVStream *st = s->streams[n];

View File

@ -65,7 +65,7 @@ static int dtshd_read_header(AVFormatContext *s)
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_DTS;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
for (;;) {
chunk_type = avio_rb64(pb);

View File

@ -508,6 +508,7 @@ static void dump_stream_format(const AVFormatContext *ic, int i,
char buf[256];
int flags = (is_output ? ic->oformat->flags : ic->iformat->flags);
const AVStream *st = ic->streams[i];
const FFStream *const sti = cffstream(st);
const AVDictionaryEntry *lang = av_dict_get(st->metadata, "language", NULL, 0);
const char *separator = ic->dump_separator;
AVCodecContext *avctx;
@ -524,12 +525,12 @@ static void dump_stream_format(const AVFormatContext *ic, int i,
}
// Fields which are missing from AVCodecParameters need to be taken from the AVCodecContext
avctx->properties = st->internal->avctx->properties;
avctx->codec = st->internal->avctx->codec;
avctx->qmin = st->internal->avctx->qmin;
avctx->qmax = st->internal->avctx->qmax;
avctx->coded_width = st->internal->avctx->coded_width;
avctx->coded_height = st->internal->avctx->coded_height;
avctx->properties = sti->avctx->properties;
avctx->codec = sti->avctx->codec;
avctx->qmin = sti->avctx->qmin;
avctx->qmax = sti->avctx->qmax;
avctx->coded_width = sti->avctx->coded_width;
avctx->coded_height = sti->avctx->coded_height;
if (separator)
av_opt_set(avctx, "dump_separator", separator, 0);
@ -544,7 +545,7 @@ static void dump_stream_format(const AVFormatContext *ic, int i,
av_log(NULL, AV_LOG_INFO, "[0x%x]", st->id);
if (lang)
av_log(NULL, AV_LOG_INFO, "(%s)", lang->value);
av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", st->internal->codec_info_nb_frames,
av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", sti->codec_info_nb_frames,
st->time_base.num, st->time_base.den);
av_log(NULL, AV_LOG_INFO, ": %s", buf);

View File

@ -503,7 +503,7 @@ static int init_video_stream(AVFormatContext *s, VideoProperties *video)
st->codecpar->codec_id = video->codec;
// parsing is necessary to make FFmpeg generate correct timestamps
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG2VIDEO)
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
st->codecpar->codec_tag = 0; /* no fourcc */
st->codecpar->width = video->width;
st->codecpar->height = video->height;

View File

@ -139,7 +139,7 @@ static int fifo_thread_write_header(FifoThreadContext *ctx)
}
for (i = 0;i < avf2->nb_streams; i++)
avf2->streams[i]->internal->cur_dts = 0;
ffstream(avf2->streams[i])->cur_dts = 0;
ret = avformat_write_header(avf2, &format_options);
if (!ret)

View File

@ -37,11 +37,10 @@ typedef struct FLACDecContext {
static void reset_index_position(int64_t metadata_head_size, AVStream *st)
{
FFStream *const sti = ffstream(st);
/* the real seek index offset should be the size of metadata blocks with the offset in the frame blocks */
int i;
for(i=0; i<st->internal->nb_index_entries; i++) {
st->internal->index_entries[i].pos += metadata_head_size;
}
for (int i = 0; i < sti->nb_index_entries; i++)
sti->index_entries[i].pos += metadata_head_size;
}
static int flac_read_header(AVFormatContext *s)
@ -55,7 +54,7 @@ static int flac_read_header(AVFormatContext *s)
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_FLAC;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
/* the parameters will be extracted from the compressed bitstream */
/* if fLaC marker is not found, assume there is no header */
@ -287,7 +286,7 @@ static av_unused int64_t flac_read_timestamp(AVFormatContext *s, int stream_inde
av_assert1(!pkt->size);
}
}
av_parser_parse2(parser, st->internal->avctx,
av_parser_parse2(parser, ffstream(st)->avctx,
&data, &size, pkt->data, pkt->size,
pkt->pts, pkt->dts, *ppos);
@ -308,6 +307,8 @@ static av_unused int64_t flac_read_timestamp(AVFormatContext *s, int stream_inde
}
static int flac_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int flags) {
AVStream *const st = s->streams[0];
FFStream *const sti = ffstream(st);
int index;
int64_t pos;
AVIndexEntry e;
@ -317,11 +318,11 @@ static int flac_seek(AVFormatContext *s, int stream_index, int64_t timestamp, in
return -1;
}
index = av_index_search_timestamp(s->streams[0], timestamp, flags);
if(index<0 || index >= s->streams[0]->internal->nb_index_entries)
index = av_index_search_timestamp(st, timestamp, flags);
if (index < 0 || index >= sti->nb_index_entries)
return -1;
e = s->streams[0]->internal->index_entries[index];
e = sti->index_entries[index];
pos = avio_seek(s->pb, e.pos, SEEK_SET);
if (pos >= 0) {
return 0;

View File

@ -265,10 +265,11 @@ static int flic_read_seek(AVFormatContext *s, int stream_index,
{
FlicDemuxContext *flic = s->priv_data;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int64_t pos, ts;
int index;
if (!st->internal->index_entries || stream_index != flic->video_stream_index)
if (!sti->index_entries || stream_index != flic->video_stream_index)
return -1;
index = av_index_search_timestamp(st, pts, flags);
@ -278,8 +279,8 @@ static int flic_read_seek(AVFormatContext *s, int stream_index,
if (index < 0)
return -1;
pos = st->internal->index_entries[index].pos;
ts = st->internal->index_entries[index].timestamp;
pos = sti->index_entries[index].pos;
ts = sti->index_entries[index].timestamp;
flic->frame_number = ts;
avio_seek(s->pb, pos, SEEK_SET);
return 0;

View File

@ -143,7 +143,7 @@ static void add_keyframes_index(AVFormatContext *s)
av_assert0(flv->last_keyframe_stream_index <= s->nb_streams);
stream = s->streams[flv->last_keyframe_stream_index];
if (stream->internal->nb_index_entries == 0) {
if (ffstream(stream)->nb_index_entries == 0) {
for (i = 0; i < flv->keyframe_count; i++) {
av_log(s, AV_LOG_TRACE, "keyframe filepositions = %"PRId64" times = %"PRId64"\n",
flv->keyframe_filepositions[i], flv->keyframe_times[i] * 1000);
@ -272,7 +272,7 @@ static void flv_set_audio_codec(AVFormatContext *s, AVStream *astream,
break;
case FLV_CODECID_MP3:
apar->codec_id = AV_CODEC_ID_MP3;
astream->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(astream)->need_parsing = AVSTREAM_PARSE_FULL;
break;
case FLV_CODECID_NELLYMOSER_8KHZ_MONO:
// in case metadata does not otherwise declare samplerate
@ -329,6 +329,7 @@ static int flv_same_video_codec(AVCodecParameters *vpar, int flags)
static int flv_set_video_codec(AVFormatContext *s, AVStream *vstream,
int flv_codecid, int read)
{
FFStream *const vstreami = ffstream(vstream);
int ret = 0;
AVCodecParameters *par = vstream->codecpar;
enum AVCodecID old_codec_id = vstream->codecpar->codec_id;
@ -363,7 +364,7 @@ static int flv_set_video_codec(AVFormatContext *s, AVStream *vstream,
break;
case FLV_CODECID_H264:
par->codec_id = AV_CODEC_ID_H264;
vstream->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
vstreami->need_parsing = AVSTREAM_PARSE_HEADERS;
ret = 3; // not 4, reading packet type will consume one byte
break;
case FLV_CODECID_MPEG4:
@ -375,7 +376,7 @@ static int flv_set_video_codec(AVFormatContext *s, AVStream *vstream,
par->codec_tag = flv_codecid;
}
if (!vstream->internal->need_context_update && par->codec_id != old_codec_id) {
if (!vstreami->need_context_update && par->codec_id != old_codec_id) {
avpriv_request_sample(s, "Changing the codec id midstream");
return AVERROR_PATCHWELCOME;
}
@ -815,7 +816,7 @@ static int flv_get_extradata(AVFormatContext *s, AVStream *st, int size)
if ((ret = ff_get_extradata(s, st->codecpar, s->pb, size)) < 0)
return ret;
st->internal->need_context_update = 1;
ffstream(st)->need_context_update = 1;
return 0;
}
@ -837,17 +838,16 @@ static int flv_queue_extradata(FLVContext *flv, AVIOContext *pb, int stream,
static void clear_index_entries(AVFormatContext *s, int64_t pos)
{
int i, j, out;
av_log(s, AV_LOG_WARNING,
"Found invalid index entries, clearing the index.\n");
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
for (unsigned i = 0; i < s->nb_streams; i++) {
FFStream *const sti = ffstream(s->streams[i]);
int out = 0;
/* Remove all index entries that point to >= pos */
out = 0;
for (j = 0; j < st->internal->nb_index_entries; j++)
if (st->internal->index_entries[j].pos < pos)
st->internal->index_entries[out++] = st->internal->index_entries[j];
st->internal->nb_index_entries = out;
for (int j = 0; j < sti->nb_index_entries; j++)
if (sti->index_entries[j].pos < pos)
sti->index_entries[out++] = sti->index_entries[j];
sti->nb_index_entries = out;
}
}

View File

@ -104,12 +104,14 @@ static int gxf_probe(const AVProbeData *p) {
static int get_sindex(AVFormatContext *s, int id, int format) {
int i;
AVStream *st = NULL;
FFStream *sti;
i = ff_find_stream_index(s, id);
if (i >= 0)
return i;
st = avformat_new_stream(s, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
st->id = id;
switch (format) {
case 3:
@ -130,13 +132,13 @@ static int get_sindex(AVFormatContext *s, int id, int format) {
case 20:
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG2VIDEO;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS; //get keyframe flag etc.
sti->need_parsing = AVSTREAM_PARSE_HEADERS; //get keyframe flag etc.
break;
case 22:
case 23:
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG1VIDEO;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS; //get keyframe flag etc.
sti->need_parsing = AVSTREAM_PARSE_HEADERS; //get keyframe flag etc.
break;
case 9:
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
@ -169,7 +171,7 @@ static int get_sindex(AVFormatContext *s, int id, int format) {
case 29: /* AVCHD */
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_H264;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
break;
// timecode tracks:
case 7:
@ -567,6 +569,7 @@ static int gxf_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int
uint64_t pos;
uint64_t maxlen = 100 * 1024 * 1024;
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
int64_t start_time = s->streams[stream_index]->start_time;
int64_t found;
int idx;
@ -575,9 +578,9 @@ static int gxf_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int
AVSEEK_FLAG_ANY | AVSEEK_FLAG_BACKWARD);
if (idx < 0)
return -1;
pos = st->internal->index_entries[idx].pos;
if (idx < st->internal->nb_index_entries - 2)
maxlen = st->internal->index_entries[idx + 2].pos - pos;
pos = sti->index_entries[idx].pos;
if (idx < sti->nb_index_entries - 2)
maxlen = sti->index_entries[idx + 2].pos - pos;
maxlen = FFMAX(maxlen, 200 * 1024);
res = avio_seek(s->pb, pos, SEEK_SET);
if (res < 0)

View File

@ -500,15 +500,16 @@ static int hds_write_packet(AVFormatContext *s, AVPacket *pkt)
{
HDSContext *c = s->priv_data;
AVStream *st = s->streams[pkt->stream_index];
FFStream *const sti = ffstream(st);
OutputStream *os = &c->streams[s->streams[pkt->stream_index]->id];
int64_t end_dts = os->fragment_index * (int64_t)c->min_frag_duration;
int ret;
if (st->internal->first_dts == AV_NOPTS_VALUE)
st->internal->first_dts = pkt->dts;
if (sti->first_dts == AV_NOPTS_VALUE)
sti->first_dts = pkt->dts;
if ((!os->has_video || st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) &&
av_compare_ts(pkt->dts - st->internal->first_dts, st->time_base,
av_compare_ts(pkt->dts - sti->first_dts, st->time_base,
end_dts, AV_TIME_BASE_Q) >= 0 &&
pkt->flags & AV_PKT_FLAG_KEY && os->packets_written) {

View File

@ -1794,7 +1794,7 @@ static int set_stream_info_from_input_stream(AVStream *st, struct playlist *pls,
memcpy(dst_data, sd_src->data, sd_src->size);
}
st->internal->need_context_update = 1;
ffstream(st)->need_context_update = 1;
return 0;
}

View File

@ -188,23 +188,26 @@ static int ifv_read_header(AVFormatContext *s)
static int ifv_read_packet(AVFormatContext *s, AVPacket *pkt)
{
IFVContext *ifv = s->priv_data;
AVStream *st;
AVIndexEntry *ev, *ea, *e_next;
int ret;
ev = ea = e_next = NULL;
if (ifv->next_video_index < ifv->total_vframes) {
st = s->streams[ifv->video_stream_index];
if (ifv->next_video_index < st->internal->nb_index_entries)
e_next = ev = &st->internal->index_entries[ifv->next_video_index];
AVStream *const st = s->streams[ifv->video_stream_index];
FFStream *const sti = ffstream(st);
if (ifv->next_video_index < sti->nb_index_entries)
e_next = ev = &sti->index_entries[ifv->next_video_index];
}
if (ifv->is_audio_present &&
ifv->next_audio_index < ifv->total_aframes) {
st = s->streams[ifv->audio_stream_index];
if (ifv->next_audio_index < st->internal->nb_index_entries) {
ea = &st->internal->index_entries[ifv->next_audio_index];
AVStream *const st = s->streams[ifv->audio_stream_index];
FFStream *const sti = ffstream(st);
if (ifv->next_audio_index < sti->nb_index_entries) {
ea = &sti->index_entries[ifv->next_audio_index];
if (!ev || ea->timestamp < ev->timestamp)
e_next = ea;
}

View File

@ -209,7 +209,7 @@ int ff_img_read_header(AVFormatContext *s1)
s->is_pipe = 0;
else {
s->is_pipe = 1;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
}
if (s->ts_from_file == 2) {
@ -482,7 +482,7 @@ int ff_img_read_packet(AVFormatContext *s1, AVPacket *pkt)
return AVERROR_EOF;
if (s->frame_size > 0) {
size[0] = s->frame_size;
} else if (!s1->streams[0]->internal->parser) {
} else if (!ffstream(s1->streams[0])->parser) {
size[0] = avio_size(s1->pb);
} else {
size[0] = 4096;
@ -590,7 +590,7 @@ static int img_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
int index = av_index_search_timestamp(st, timestamp, flags);
if(index < 0)
return -1;
s1->img_number = st->internal->index_entries[index].pos;
s1->img_number = ffstream(st)->index_entries[index].pos;
return 0;
}

View File

@ -185,7 +185,12 @@ static av_always_inline FFFormatContext *ffformatcontext(AVFormatContext *s)
return (FFFormatContext*)s;
}
struct AVStreamInternal {
typedef struct FFStream {
/**
* The public context.
*/
AVStream pub;
/**
* Set to 1 if the codec allows reordering, so pts can be different
* from dts.
@ -416,7 +421,17 @@ struct AVStreamInternal {
*/
int64_t first_dts;
int64_t cur_dts;
};
} FFStream;
static av_always_inline FFStream *ffstream(AVStream *st)
{
return (FFStream*)st;
}
static av_always_inline const FFStream *cffstream(const AVStream *st)
{
return (FFStream*)st;
}
void avpriv_stream_set_need_parsing(AVStream *st, enum AVStreamParseType type);

View File

@ -62,7 +62,7 @@ static int ipu_read_header(AVFormatContext *s)
st->start_time = 0;
st->duration =
st->nb_frames = avio_rl32(pb);
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 1, 25);
return 0;

View File

@ -47,7 +47,7 @@ static int read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG4;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_set_pts_info(st, 64, 1, 90000);
return 0;

View File

@ -56,7 +56,7 @@ static int read_header(AVFormatContext *s)
st->duration = avio_rl32(s->pb);
avio_skip(s->pb, 4); // unused
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
if (!time_base.den || !time_base.num) {
av_log(s, AV_LOG_ERROR, "Invalid frame rate\n");

View File

@ -74,9 +74,9 @@ static int read_header(AVFormatContext *s)
JVDemuxContext *jv = s->priv_data;
AVIOContext *pb = s->pb;
AVStream *vst, *ast;
FFStream *asti;
int64_t audio_pts = 0;
int64_t offset;
int i;
avio_skip(pb, 80);
@ -84,6 +84,7 @@ static int read_header(AVFormatContext *s)
vst = avformat_new_stream(s, NULL);
if (!ast || !vst)
return AVERROR(ENOMEM);
asti = ffstream(ast);
vst->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
vst->codecpar->codec_id = AV_CODEC_ID_JV;
@ -92,7 +93,7 @@ static int read_header(AVFormatContext *s)
vst->codecpar->height = avio_rl16(pb);
vst->duration =
vst->nb_frames =
ast->internal->nb_index_entries = avio_rl16(pb);
asti->nb_index_entries = avio_rl16(pb);
avpriv_set_pts_info(vst, 64, avio_rl16(pb), 1000);
avio_skip(pb, 4);
@ -107,17 +108,17 @@ static int read_header(AVFormatContext *s)
avio_skip(pb, 10);
ast->internal->index_entries = av_malloc(ast->internal->nb_index_entries *
sizeof(*ast->internal->index_entries));
if (!ast->internal->index_entries)
asti->index_entries = av_malloc(asti->nb_index_entries *
sizeof(*asti->index_entries));
if (!asti->index_entries)
return AVERROR(ENOMEM);
jv->frames = av_malloc(ast->internal->nb_index_entries * sizeof(JVFrame));
jv->frames = av_malloc(asti->nb_index_entries * sizeof(*jv->frames));
if (!jv->frames)
return AVERROR(ENOMEM);
offset = 0x68 + ast->internal->nb_index_entries * 16;
for (i = 0; i < ast->internal->nb_index_entries; i++) {
AVIndexEntry *e = ast->internal->index_entries + i;
offset = 0x68 + asti->nb_index_entries * 16;
for (int i = 0; i < asti->nb_index_entries; i++) {
AVIndexEntry *e = asti->index_entries + i;
JVFrame *jvf = jv->frames + i;
/* total frame size including audio, video, palette data and padding */
@ -162,10 +163,11 @@ static int read_packet(AVFormatContext *s, AVPacket *pkt)
JVDemuxContext *jv = s->priv_data;
AVIOContext *pb = s->pb;
AVStream *ast = s->streams[0];
FFStream *const asti = ffstream(ast);
int ret;
while (!avio_feof(s->pb) && jv->pts < ast->internal->nb_index_entries) {
const AVIndexEntry *e = ast->internal->index_entries + jv->pts;
while (!avio_feof(s->pb) && jv->pts < asti->nb_index_entries) {
const AVIndexEntry *const e = asti->index_entries + jv->pts;
const JVFrame *jvf = jv->frames + jv->pts;
switch (jv->state) {
@ -222,6 +224,7 @@ static int read_seek(AVFormatContext *s, int stream_index,
{
JVDemuxContext *jv = s->priv_data;
AVStream *ast = s->streams[0];
FFStream *const asti = ffstream(ast);
int i;
if (flags & (AVSEEK_FLAG_BYTE | AVSEEK_FLAG_FRAME))
@ -238,9 +241,9 @@ static int read_seek(AVFormatContext *s, int stream_index,
return 0;
}
if (i < 0 || i >= ast->internal->nb_index_entries)
if (i < 0 || i >= asti->nb_index_entries)
return 0;
if (avio_seek(s->pb, ast->internal->index_entries[i].pos, SEEK_SET) < 0)
if (avio_seek(s->pb, asti->index_entries[i].pos, SEEK_SET) < 0)
return -1;
jv->state = JV_AUDIO;

View File

@ -67,14 +67,14 @@ static int lmlm4_read_header(AVFormatContext *s)
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG4;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
avpriv_set_pts_info(st, 64, 1001, 30000);
if (!(st = avformat_new_stream(s, NULL)))
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_MP2;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
/* the parameters will be extracted from the compressed bitstream */
return 0;

View File

@ -75,7 +75,7 @@ static int loas_read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = s->iformat->raw_codec_id;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
//LCM of all possible AAC sample rates
avpriv_set_pts_info(st, 64, 1, 28224000);

View File

@ -262,7 +262,7 @@ static int lxf_read_header(AVFormatContext *s)
st->codecpar->bit_rate = 1000000 * ((video_params >> 14) & 0xFF);
st->codecpar->codec_tag = video_params & 0xF;
st->codecpar->codec_id = ff_codec_get_id(lxf_tags, st->codecpar->codec_tag);
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
av_log(s, AV_LOG_DEBUG, "record: %x = %i-%02i-%02i\n",
record_date, 1900 + (record_date & 0x7F), (record_date >> 7) & 0xF,

View File

@ -2338,7 +2338,6 @@ static int matroska_parse_tracks(AVFormatContext *s)
{
MatroskaDemuxContext *matroska = s->priv_data;
MatroskaTrack *tracks = matroska->tracks.elem;
AVStream *st;
int i, j, ret;
int k;
@ -2352,6 +2351,8 @@ static int matroska_parse_tracks(AVFormatContext *s)
int extradata_offset = 0;
uint32_t fourcc = 0;
FFIOContext b;
AVStream *st;
FFStream *sti;
char* key_id_base64 = NULL;
int bit_depth = -1;
@ -2486,6 +2487,7 @@ static int matroska_parse_tracks(AVFormatContext *s)
av_free(key_id_base64);
return AVERROR(ENOMEM);
}
sti = ffstream(st);
if (key_id_base64) {
/* export encryption key id as base64 metadata tag */
@ -2802,7 +2804,7 @@ static int matroska_parse_tracks(AVFormatContext *s)
255);
}
if (st->codecpar->codec_id != AV_CODEC_ID_HEVC)
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
if (track->default_duration) {
int div = track->default_duration <= INT64_MAX ? 1 : 2;
@ -2861,9 +2863,9 @@ static int matroska_parse_tracks(AVFormatContext *s)
if (st->codecpar->codec_id == AV_CODEC_ID_MP3 ||
st->codecpar->codec_id == AV_CODEC_ID_MLP ||
st->codecpar->codec_id == AV_CODEC_ID_TRUEHD)
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
else if (st->codecpar->codec_id != AV_CODEC_ID_AAC)
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
if (track->codec_delay > 0) {
st->codecpar->initial_padding = av_rescale_q(track->codec_delay,
(AVRational){1, 1000000000},
@ -3652,7 +3654,7 @@ static int matroska_parse_block(MatroskaDemuxContext *matroska, AVBufferRef *buf
return res;
if (is_keyframe)
matroska->skip_to_keyframe = 0;
else if (!st->internal->skip_to_keyframe) {
else if (!ffstream(st)->skip_to_keyframe) {
av_log(matroska->ctx, AV_LOG_ERROR, "File is broken, keyframes not correctly marked!\n");
matroska->skip_to_keyframe = 0;
}
@ -3814,6 +3816,7 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index,
MatroskaDemuxContext *matroska = s->priv_data;
MatroskaTrack *tracks = NULL;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int i, index;
/* Parse the CUES now since we need the index data to seek. */
@ -3822,13 +3825,15 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index,
matroska_parse_cues(matroska);
}
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
goto err;
timestamp = FFMAX(timestamp, st->internal->index_entries[0].timestamp);
timestamp = FFMAX(timestamp, sti->index_entries[0].timestamp);
if ((index = av_index_search_timestamp(st, timestamp, flags)) < 0 || index == st->internal->nb_index_entries - 1) {
matroska_reset_status(matroska, 0, st->internal->index_entries[st->internal->nb_index_entries - 1].pos);
while ((index = av_index_search_timestamp(st, timestamp, flags)) < 0 || index == st->internal->nb_index_entries - 1) {
if ((index = av_index_search_timestamp(st, timestamp, flags)) < 0 ||
index == sti->nb_index_entries - 1) {
matroska_reset_status(matroska, 0, sti->index_entries[sti->nb_index_entries - 1].pos);
while ((index = av_index_search_timestamp(st, timestamp, flags)) < 0 ||
index == sti->nb_index_entries - 1) {
matroska_clear_queue(matroska);
if (matroska_parse_cluster(matroska) < 0)
break;
@ -3836,7 +3841,8 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index,
}
matroska_clear_queue(matroska);
if (index < 0 || (matroska->cues_parsing_deferred < 0 && index == st->internal->nb_index_entries - 1))
if (index < 0 || (matroska->cues_parsing_deferred < 0 &&
index == sti->nb_index_entries - 1))
goto err;
tracks = matroska->tracks.elem;
@ -3848,17 +3854,17 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index,
}
/* We seek to a level 1 element, so set the appropriate status. */
matroska_reset_status(matroska, 0, st->internal->index_entries[index].pos);
matroska_reset_status(matroska, 0, sti->index_entries[index].pos);
if (flags & AVSEEK_FLAG_ANY) {
st->internal->skip_to_keyframe = 0;
sti->skip_to_keyframe = 0;
matroska->skip_to_timecode = timestamp;
} else {
st->internal->skip_to_keyframe = 1;
matroska->skip_to_timecode = st->internal->index_entries[index].timestamp;
sti->skip_to_keyframe = 1;
matroska->skip_to_timecode = sti->index_entries[index].timestamp;
}
matroska->skip_to_keyframe = 1;
matroska->done = 0;
avpriv_update_cur_dts(s, st, st->internal->index_entries[index].timestamp);
avpriv_update_cur_dts(s, st, sti->index_entries[index].timestamp);
return 0;
err:
// slightly hackish but allows proper fallback to
@ -3866,7 +3872,7 @@ err:
matroska_reset_status(matroska, 0, -1);
matroska->resync_pos = -1;
matroska_clear_queue(matroska);
st->internal->skip_to_keyframe =
sti->skip_to_keyframe =
matroska->skip_to_keyframe = 0;
matroska->done = 0;
return -1;
@ -3902,10 +3908,12 @@ typedef struct {
*/
static CueDesc get_cue_desc(AVFormatContext *s, int64_t ts, int64_t cues_start) {
MatroskaDemuxContext *matroska = s->priv_data;
FFStream *const sti = ffstream(s->streams[0]);
AVIndexEntry *const index_entries = sti->index_entries;
int nb_index_entries = sti->nb_index_entries;
CueDesc cue_desc;
int i;
int nb_index_entries = s->streams[0]->internal->nb_index_entries;
AVIndexEntry *index_entries = s->streams[0]->internal->index_entries;
if (ts >= matroska->duration * matroska->time_scale) return (CueDesc) {-1, -1, -1, -1};
for (i = 1; i < nb_index_entries; i++) {
if (index_entries[i - 1].timestamp * matroska->time_scale <= ts &&
@ -3932,14 +3940,20 @@ static CueDesc get_cue_desc(AVFormatContext *s, int64_t ts, int64_t cues_start)
static int webm_clusters_start_with_keyframe(AVFormatContext *s)
{
MatroskaDemuxContext *matroska = s->priv_data;
AVStream *const st = s->streams[0];
FFStream *const sti = ffstream(st);
uint32_t id = matroska->current_id;
int64_t cluster_pos, before_pos;
int index, rv = 1;
if (s->streams[0]->internal->nb_index_entries <= 0) return 0;
if (sti->nb_index_entries <= 0)
return 0;
// seek to the first cluster using cues.
index = av_index_search_timestamp(s->streams[0], 0, 0);
if (index < 0) return 0;
cluster_pos = s->streams[0]->internal->index_entries[index].pos;
index = av_index_search_timestamp(st, 0, 0);
if (index < 0)
return 0;
cluster_pos = sti->index_entries[index].pos;
before_pos = avio_tell(s->pb);
while (1) {
uint64_t cluster_id, cluster_length;
@ -4060,12 +4074,12 @@ static int64_t webm_dash_manifest_compute_bandwidth(AVFormatContext *s, int64_t
{
MatroskaDemuxContext *matroska = s->priv_data;
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
double bandwidth = 0.0;
int i;
for (i = 0; i < st->internal->nb_index_entries; i++) {
for (int i = 0; i < sti->nb_index_entries; i++) {
int64_t prebuffer_ns = 1000000000;
int64_t time_ns = st->internal->index_entries[i].timestamp * matroska->time_scale;
int64_t time_ns = sti->index_entries[i].timestamp * matroska->time_scale;
double nano_seconds_per_second = 1000000000.0;
int64_t prebuffered_ns = time_ns + prebuffer_ns;
double prebuffer_bytes = 0.0;
@ -4149,6 +4163,7 @@ static int webm_dash_manifest_cues(AVFormatContext *s, int64_t init_range)
EbmlList *seekhead_list = &matroska->seekhead;
MatroskaSeekhead *seekhead = seekhead_list->elem;
AVStream *const st = s->streams[0];
FFStream *const sti = ffstream(st);
AVBPrint bprint;
char *buf;
int64_t cues_start = -1, cues_end = -1, before_pos, bandwidth;
@ -4184,7 +4199,7 @@ static int webm_dash_manifest_cues(AVFormatContext *s, int64_t init_range)
// parse the cues
matroska_parse_cues(matroska);
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
return AVERROR_INVALIDDATA;
// cues start
@ -4209,8 +4224,8 @@ static int webm_dash_manifest_cues(AVFormatContext *s, int64_t init_range)
// Store cue point timestamps as a comma separated list
// for checking subsegment alignment in the muxer.
av_bprint_init(&bprint, 0, AV_BPRINT_SIZE_UNLIMITED);
for (int i = 0; i < st->internal->nb_index_entries; i++)
av_bprintf(&bprint, "%" PRId64",", st->internal->index_entries[i].timestamp);
for (int i = 0; i < sti->nb_index_entries; i++)
av_bprintf(&bprint, "%" PRId64",", sti->index_entries[i].timestamp);
if (!av_bprint_is_complete(&bprint)) {
av_bprint_finalize(&bprint, NULL);
return AVERROR(ENOMEM);

View File

@ -50,7 +50,7 @@ static int read_header(AVFormatContext *s)
if (!st)
return AVERROR(ENOMEM);
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
st->start_time = 0;
st->nb_frames =
st->duration = avio_rb32(pb);

View File

@ -117,6 +117,7 @@ static void read_uint64(AVFormatContext *avctx, AVIOContext *pb, const char *tag
static int scan_file(AVFormatContext *avctx, AVStream *vst, AVStream *ast, int file)
{
FFStream *const vsti = ffstream(vst), *const asti = ffstream(ast);
MlvContext *mlv = avctx->priv_data;
AVIOContext *pb = mlv->pb[file];
int ret;
@ -189,12 +190,14 @@ static int scan_file(AVFormatContext *avctx, AVStream *vst, AVStream *ast, int f
}
} else if (vst && type == MKTAG('V', 'I', 'D', 'F') && size >= 4) {
uint64_t pts = avio_rl32(pb);
ff_add_index_entry(&vst->internal->index_entries, &vst->internal->nb_index_entries, &vst->internal->index_entries_allocated_size,
ff_add_index_entry(&vsti->index_entries, &vsti->nb_index_entries,
&vsti->index_entries_allocated_size,
avio_tell(pb) - 20, pts, file, 0, AVINDEX_KEYFRAME);
size -= 4;
} else if (ast && type == MKTAG('A', 'U', 'D', 'F') && size >= 4) {
uint64_t pts = avio_rl32(pb);
ff_add_index_entry(&ast->internal->index_entries, &ast->internal->nb_index_entries, &ast->internal->index_entries_allocated_size,
ff_add_index_entry(&asti->index_entries, &asti->nb_index_entries,
&asti->index_entries_allocated_size,
avio_tell(pb) - 20, pts, file, 0, AVINDEX_KEYFRAME);
size -= 4;
} else if (vst && type == MKTAG('W','B','A','L') && size >= 28) {
@ -257,6 +260,7 @@ static int read_header(AVFormatContext *avctx)
MlvContext *mlv = avctx->priv_data;
AVIOContext *pb = avctx->pb;
AVStream *vst = NULL, *ast = NULL;
FFStream *vsti = NULL, *asti = NULL;
int size, ret;
unsigned nb_video_frames, nb_audio_frames;
uint64_t guid;
@ -285,6 +289,8 @@ static int read_header(AVFormatContext *avctx)
vst = avformat_new_stream(avctx, NULL);
if (!vst)
return AVERROR(ENOMEM);
vsti = ffstream(vst);
vst->id = 0;
vst->nb_frames = nb_video_frames;
if ((mlv->class[0] & (MLV_CLASS_FLAG_DELTA|MLV_CLASS_FLAG_LZMA)))
@ -316,6 +322,7 @@ static int read_header(AVFormatContext *avctx)
ast = avformat_new_stream(avctx, NULL);
if (!ast)
return AVERROR(ENOMEM);
asti = ffstream(ast);
ast->id = 1;
ast->nb_frames = nb_audio_frames;
if ((mlv->class[1] & MLV_CLASS_FLAG_LZMA))
@ -372,21 +379,21 @@ static int read_header(AVFormatContext *avctx)
}
if (vst)
vst->duration = vst->internal->nb_index_entries;
vst->duration = vsti->nb_index_entries;
if (ast)
ast->duration = ast->internal->nb_index_entries;
ast->duration = asti->nb_index_entries;
if ((vst && !vst->internal->nb_index_entries) || (ast && !ast->internal->nb_index_entries)) {
if ((vst && !vsti->nb_index_entries) || (ast && !asti->nb_index_entries)) {
av_log(avctx, AV_LOG_ERROR, "no index entries found\n");
return AVERROR_INVALIDDATA;
}
if (vst && ast)
avio_seek(pb, FFMIN(vst->internal->index_entries[0].pos, ast->internal->index_entries[0].pos), SEEK_SET);
avio_seek(pb, FFMIN(vsti->index_entries[0].pos, asti->index_entries[0].pos), SEEK_SET);
else if (vst)
avio_seek(pb, vst->internal->index_entries[0].pos, SEEK_SET);
avio_seek(pb, vsti->index_entries[0].pos, SEEK_SET);
else if (ast)
avio_seek(pb, ast->internal->index_entries[0].pos, SEEK_SET);
avio_seek(pb, asti->index_entries[0].pos, SEEK_SET);
return 0;
}
@ -396,6 +403,7 @@ static int read_packet(AVFormatContext *avctx, AVPacket *pkt)
MlvContext *mlv = avctx->priv_data;
AVIOContext *pb;
AVStream *st;
FFStream *sti;
int index, ret;
unsigned int size, space;
@ -403,6 +411,7 @@ static int read_packet(AVFormatContext *avctx, AVPacket *pkt)
return AVERROR_EOF;
st = avctx->streams[mlv->stream_index];
sti = ffstream(st);
if (mlv->pts >= st->duration)
return AVERROR_EOF;
@ -412,12 +421,12 @@ static int read_packet(AVFormatContext *avctx, AVPacket *pkt)
return AVERROR(EIO);
}
pb = mlv->pb[st->internal->index_entries[index].size];
pb = mlv->pb[sti->index_entries[index].size];
if (!pb) {
ret = FFERROR_REDO;
goto next_packet;
}
avio_seek(pb, st->internal->index_entries[index].pos, SEEK_SET);
avio_seek(pb, sti->index_entries[index].pos, SEEK_SET);
avio_skip(pb, 4); // blockType
size = avio_rl32(pb);

View File

@ -1767,7 +1767,10 @@ static int mov_read_targa_y216(MOVContext *c, AVIOContext *pb, MOVAtom atom)
static int mov_read_ares(MOVContext *c, AVIOContext *pb, MOVAtom atom)
{
if (c->fc->nb_streams >= 1) {
AVCodecParameters *par = c->fc->streams[c->fc->nb_streams-1]->codecpar;
AVStream *const st = c->fc->streams[c->fc->nb_streams - 1];
FFStream *const sti = ffstream(st);
AVCodecParameters *par = st->codecpar;
if (par->codec_tag == MKTAG('A', 'V', 'i', 'n') &&
par->codec_id == AV_CODEC_ID_H264 &&
atom.size > 11) {
@ -1794,8 +1797,7 @@ static int mov_read_ares(MOVContext *c, AVIOContext *pb, MOVAtom atom)
return 0;
den *= 2;
case 1:
c->fc->streams[c->fc->nb_streams-1]->internal->display_aspect_ratio.num = num;
c->fc->streams[c->fc->nb_streams-1]->internal->display_aspect_ratio.den = den;
sti->display_aspect_ratio = (AVRational){ num, den };
default:
return 0;
}
@ -2198,7 +2200,7 @@ static void mov_parse_stsd_audio(MOVContext *c, AVIOContext *pb,
switch (st->codecpar->codec_id) {
case AV_CODEC_ID_MP2:
case AV_CODEC_ID_MP3:
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
break;
}
}
@ -2368,6 +2370,8 @@ static int mov_parse_stsd_data(MOVContext *c, AVIOContext *pb,
static int mov_finalize_stsd_codec(MOVContext *c, AVIOContext *pb,
AVStream *st, MOVStreamContext *sc)
{
FFStream *const sti = ffstream(st);
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO &&
!st->codecpar->sample_rate && sc->time_scale > 1)
st->codecpar->sample_rate = sc->time_scale;
@ -2436,10 +2440,10 @@ static int mov_finalize_stsd_codec(MOVContext *c, AVIOContext *pb,
case AV_CODEC_ID_VC1:
case AV_CODEC_ID_VP8:
case AV_CODEC_ID_VP9:
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
break;
case AV_CODEC_ID_AV1:
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
break;
default:
break;
@ -2768,12 +2772,14 @@ static int mov_read_stps(MOVContext *c, AVIOContext *pb, MOVAtom atom)
static int mov_read_stss(MOVContext *c, AVIOContext *pb, MOVAtom atom)
{
AVStream *st;
FFStream *sti;
MOVStreamContext *sc;
unsigned int i, entries;
if (c->fc->nb_streams < 1)
return 0;
st = c->fc->streams[c->fc->nb_streams-1];
sti = ffstream(st);
sc = st->priv_data;
avio_r8(pb); /* version */
@ -2785,8 +2791,8 @@ static int mov_read_stss(MOVContext *c, AVIOContext *pb, MOVAtom atom)
if (!entries) {
sc->keyframe_absent = 1;
if (!st->internal->need_parsing && st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
if (!sti->need_parsing && st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
return 0;
}
if (sc->keyframes)
@ -3191,8 +3197,9 @@ static int find_prev_closest_index(AVStream *st,
int64_t* ctts_sample)
{
MOVStreamContext *msc = st->priv_data;
AVIndexEntry *e_keep = st->internal->index_entries;
int nb_keep = st->internal->nb_index_entries;
FFStream *const sti = ffstream(st);
AVIndexEntry *e_keep = sti->index_entries;
int nb_keep = sti->nb_index_entries;
int64_t i = 0;
int64_t index_ctts_count;
@ -3205,8 +3212,8 @@ static int find_prev_closest_index(AVStream *st,
timestamp_pts -= msc->dts_shift;
}
st->internal->index_entries = e_old;
st->internal->nb_index_entries = nb_old;
sti->index_entries = e_old;
sti->nb_index_entries = nb_old;
*index = av_index_search_timestamp(st, timestamp_pts, flag | AVSEEK_FLAG_BACKWARD);
// Keep going backwards in the index entries until the timestamp is the same.
@ -3259,14 +3266,14 @@ static int find_prev_closest_index(AVStream *st,
}
/* restore AVStream state*/
st->internal->index_entries = e_keep;
st->internal->nb_index_entries = nb_keep;
sti->index_entries = e_keep;
sti->nb_index_entries = nb_keep;
return *index >= 0 ? 0 : -1;
}
/**
* Add index entry with the given values, to the end of st->internal->index_entries.
* Returns the new size st->internal->index_entries if successful, else returns -1.
* Add index entry with the given values, to the end of ffstream(st)->index_entries.
* Returns the new size ffstream(st)->index_entries if successful, else returns -1.
*
* This function is similar to ff_add_index_entry in libavformat/utils.c
* except that here we are always unconditionally adding an index entry to
@ -3278,29 +3285,30 @@ static int find_prev_closest_index(AVStream *st,
static int64_t add_index_entry(AVStream *st, int64_t pos, int64_t timestamp,
int size, int distance, int flags)
{
FFStream *const sti = ffstream(st);
AVIndexEntry *entries, *ie;
int64_t index = -1;
const size_t min_size_needed = (st->internal->nb_index_entries + 1) * sizeof(AVIndexEntry);
const size_t min_size_needed = (sti->nb_index_entries + 1) * sizeof(AVIndexEntry);
// Double the allocation each time, to lower memory fragmentation.
// Another difference from ff_add_index_entry function.
const size_t requested_size =
min_size_needed > st->internal->index_entries_allocated_size ?
FFMAX(min_size_needed, 2 * st->internal->index_entries_allocated_size) :
min_size_needed > sti->index_entries_allocated_size ?
FFMAX(min_size_needed, 2 * sti->index_entries_allocated_size) :
min_size_needed;
if (st->internal->nb_index_entries + 1U >= UINT_MAX / sizeof(AVIndexEntry))
if (sti->nb_index_entries + 1U >= UINT_MAX / sizeof(AVIndexEntry))
return -1;
entries = av_fast_realloc(st->internal->index_entries,
&st->internal->index_entries_allocated_size,
entries = av_fast_realloc(sti->index_entries,
&sti->index_entries_allocated_size,
requested_size);
if (!entries)
return -1;
st->internal->index_entries= entries;
sti->index_entries = entries;
index= st->internal->nb_index_entries++;
index = sti->nb_index_entries++;
ie= &entries[index];
ie->pos = pos;
@ -3318,11 +3326,12 @@ static int64_t add_index_entry(AVStream *st, int64_t pos, int64_t timestamp,
static void fix_index_entry_timestamps(AVStream* st, int end_index, int64_t end_ts,
int64_t* frame_duration_buffer,
int frame_duration_buffer_size) {
FFStream *const sti = ffstream(st);
int i = 0;
av_assert0(end_index >= 0 && end_index <= st->internal->nb_index_entries);
av_assert0(end_index >= 0 && end_index <= sti->nb_index_entries);
for (i = 0; i < frame_duration_buffer_size; i++) {
end_ts -= frame_duration_buffer[frame_duration_buffer_size - 1 - i];
st->internal->index_entries[end_index - 1 - i].timestamp = end_ts;
sti->index_entries[end_index - 1 - i].timestamp = end_ts;
}
}
@ -3361,7 +3370,7 @@ static int64_t add_ctts_entry(MOVStts** ctts_data, unsigned int* ctts_count, uns
static void mov_estimate_video_delay(MOVContext *c, AVStream* st)
{
MOVStreamContext *msc = st->priv_data;
int ind;
FFStream *const sti = ffstream(st);
int ctts_ind = 0;
int ctts_sample = 0;
int64_t pts_buf[MAX_REORDER_DELAY + 1]; // Circular buffer to sort pts.
@ -3374,14 +3383,14 @@ static void mov_estimate_video_delay(MOVContext *c, AVStream* st)
if (st->codecpar->video_delay <= 0 && msc->ctts_data &&
st->codecpar->codec_id == AV_CODEC_ID_H264) {
st->codecpar->video_delay = 0;
for (ind = 0; ind < st->internal->nb_index_entries && ctts_ind < msc->ctts_count; ++ind) {
for (int ind = 0; ind < sti->nb_index_entries && ctts_ind < msc->ctts_count; ++ind) {
// Point j to the last elem of the buffer and insert the current pts there.
j = buf_start;
buf_start = (buf_start + 1);
if (buf_start == MAX_REORDER_DELAY + 1)
buf_start = 0;
pts_buf[j] = st->internal->index_entries[ind].timestamp + msc->ctts_data[ctts_ind].duration;
pts_buf[j] = sti->index_entries[ind].timestamp + msc->ctts_data[ctts_ind].duration;
// The timestamps that are already in the sorted buffer, and are greater than the
// current pts, are exactly the timestamps that need to be buffered to output PTS
@ -3461,7 +3470,7 @@ static void mov_current_sample_set(MOVStreamContext *sc, int current_sample)
}
/**
* Fix st->internal->index_entries, so that it contains only the entries (and the entries
* Fix ffstream(st)->index_entries, so that it contains only the entries (and the entries
* which are needed to decode them) that fall in the edit list time ranges.
* Also fixes the timestamps of the index entries to match the timeline
* specified the edit lists.
@ -3469,8 +3478,9 @@ static void mov_current_sample_set(MOVStreamContext *sc, int current_sample)
static void mov_fix_index(MOVContext *mov, AVStream *st)
{
MOVStreamContext *msc = st->priv_data;
AVIndexEntry *e_old = st->internal->index_entries;
int nb_old = st->internal->nb_index_entries;
FFStream *const sti = ffstream(st);
AVIndexEntry *e_old = sti->index_entries;
int nb_old = sti->nb_index_entries;
const AVIndexEntry *e_old_end = e_old + nb_old;
const AVIndexEntry *current = NULL;
MOVStts *ctts_data_old = msc->ctts_data;
@ -3497,7 +3507,6 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
int first_non_zero_audio_edit = -1;
int packet_skip_samples = 0;
MOVIndexRange *current_index_range;
int i;
int found_keyframe_after_edit = 0;
int found_non_empty_edit = 0;
@ -3515,9 +3524,9 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
current_index_range = msc->index_ranges - 1;
// Clean AVStream from traces of old index
st->internal->index_entries = NULL;
st->internal->index_entries_allocated_size = 0;
st->internal->nb_index_entries = 0;
sti->index_entries = NULL;
sti->index_entries_allocated_size = 0;
sti->nb_index_entries = 0;
// Clean ctts fields of MOVStreamContext
msc->ctts_data = NULL;
@ -3562,7 +3571,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
}
if (first_non_zero_audio_edit > 0)
st->internal->skip_samples = msc->start_pad = 0;
sti->skip_samples = msc->start_pad = 0;
}
// While reordering frame index according to edit list we must handle properly
@ -3637,7 +3646,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
curr_cts < edit_list_media_time && curr_cts + frame_duration > edit_list_media_time &&
first_non_zero_audio_edit > 0) {
packet_skip_samples = edit_list_media_time - curr_cts;
st->internal->skip_samples += packet_skip_samples;
sti->skip_samples += packet_skip_samples;
// Shift the index entry timestamp by packet_skip_samples to be correct.
edit_list_dts_counter -= packet_skip_samples;
@ -3646,7 +3655,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
// Make timestamps strictly monotonically increasing for audio, by rewriting timestamps for
// discarded packets.
if (frame_duration_buffer) {
fix_index_entry_timestamps(st, st->internal->nb_index_entries, edit_list_dts_counter,
fix_index_entry_timestamps(st, sti->nb_index_entries, edit_list_dts_counter,
frame_duration_buffer, num_discarded_begin);
av_freep(&frame_duration_buffer);
}
@ -3670,7 +3679,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
// Increment skip_samples for the first non-zero audio edit list
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO &&
first_non_zero_audio_edit > 0 && st->codecpar->codec_id != AV_CODEC_ID_VORBIS) {
st->internal->skip_samples += frame_duration;
sti->skip_samples += frame_duration;
}
}
}
@ -3685,7 +3694,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
// Make timestamps strictly monotonically increasing by rewriting timestamps for
// discarded packets.
if (frame_duration_buffer) {
fix_index_entry_timestamps(st, st->internal->nb_index_entries, edit_list_dts_counter,
fix_index_entry_timestamps(st, sti->nb_index_entries, edit_list_dts_counter,
frame_duration_buffer, num_discarded_begin);
av_freep(&frame_duration_buffer);
}
@ -3746,9 +3755,8 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
if (msc->min_corrected_pts > 0) {
av_log(mov->fc, AV_LOG_DEBUG, "Offset DTS by %"PRId64" to make first pts zero.\n", msc->min_corrected_pts);
for (i = 0; i < st->internal->nb_index_entries; ++i) {
st->internal->index_entries[i].timestamp -= msc->min_corrected_pts;
}
for (int i = 0; i < sti->nb_index_entries; ++i)
sti->index_entries[i].timestamp -= msc->min_corrected_pts;
}
}
// Start time should be equal to zero or the duration of any empty edits.
@ -3756,7 +3764,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
// Update av stream length, if it ends up shorter than the track's media duration
st->duration = FFMIN(st->duration, edit_list_dts_entry_end - start_dts);
msc->start_pad = st->internal->skip_samples;
msc->start_pad = sti->skip_samples;
// Free the old index and the old CTTS structures
av_free(e_old);
@ -3773,6 +3781,7 @@ static void mov_fix_index(MOVContext *mov, AVStream *st)
static void mov_build_index(MOVContext *mov, AVStream *st)
{
MOVStreamContext *sc = st->priv_data;
FFStream *const sti = ffstream(st);
int64_t current_offset;
int64_t current_dts = 0;
unsigned int stts_index = 0;
@ -3844,17 +3853,17 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
current_dts -= sc->dts_shift;
last_dts = current_dts;
if (!sc->sample_count || st->internal->nb_index_entries)
if (!sc->sample_count || sti->nb_index_entries)
return;
if (sc->sample_count >= UINT_MAX / sizeof(*st->internal->index_entries) - st->internal->nb_index_entries)
if (sc->sample_count >= UINT_MAX / sizeof(*sti->index_entries) - sti->nb_index_entries)
return;
if (av_reallocp_array(&st->internal->index_entries,
st->internal->nb_index_entries + sc->sample_count,
sizeof(*st->internal->index_entries)) < 0) {
st->internal->nb_index_entries = 0;
if (av_reallocp_array(&sti->index_entries,
sti->nb_index_entries + sc->sample_count,
sizeof(*sti->index_entries)) < 0) {
sti->nb_index_entries = 0;
return;
}
st->internal->index_entries_allocated_size = (st->internal->nb_index_entries + sc->sample_count) * sizeof(*st->internal->index_entries);
sti->index_entries_allocated_size = (sti->nb_index_entries + sc->sample_count) * sizeof(*sti->index_entries);
if (ctts_data_old) {
// Expand ctts entries such that we have a 1-1 mapping with samples
@ -3937,7 +3946,7 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
av_log(mov->fc, AV_LOG_ERROR, "Sample size %u is too large\n", sample_size);
return;
}
e = &st->internal->index_entries[st->internal->nb_index_entries++];
e = &sti->index_entries[sti->nb_index_entries++];
e->pos = current_offset;
e->timestamp = current_dts;
e->size = sample_size;
@ -3946,7 +3955,7 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
av_log(mov->fc, AV_LOG_TRACE, "AVIndex stream %d, sample %u, offset %"PRIx64", dts %"PRId64", "
"size %u, distance %u, keyframe %d\n", st->index, current_sample,
current_offset, current_dts, sample_size, distance, keyframe);
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && st->internal->nb_index_entries < 100)
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && sti->nb_index_entries < 100)
ff_rfps_add_frame(mov->fc, st, current_dts);
}
@ -4018,15 +4027,15 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
}
av_log(mov->fc, AV_LOG_TRACE, "chunk count %u\n", total);
if (total >= UINT_MAX / sizeof(*st->internal->index_entries) - st->internal->nb_index_entries)
if (total >= UINT_MAX / sizeof(*sti->index_entries) - sti->nb_index_entries)
return;
if (av_reallocp_array(&st->internal->index_entries,
st->internal->nb_index_entries + total,
sizeof(*st->internal->index_entries)) < 0) {
st->internal->nb_index_entries = 0;
if (av_reallocp_array(&sti->index_entries,
sti->nb_index_entries + total,
sizeof(*sti->index_entries)) < 0) {
sti->nb_index_entries = 0;
return;
}
st->internal->index_entries_allocated_size = (st->internal->nb_index_entries + total) * sizeof(*st->internal->index_entries);
sti->index_entries_allocated_size = (sti->nb_index_entries + total) * sizeof(*sti->index_entries);
// populate index
for (i = 0; i < sc->chunk_count; i++) {
@ -4061,7 +4070,7 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
}
}
if (st->internal->nb_index_entries >= total) {
if (sti->nb_index_entries >= total) {
av_log(mov->fc, AV_LOG_ERROR, "wrong chunk count %u\n", total);
return;
}
@ -4069,7 +4078,7 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
av_log(mov->fc, AV_LOG_ERROR, "Sample size %u is too large\n", size);
return;
}
e = &st->internal->index_entries[st->internal->nb_index_entries++];
e = &sti->index_entries[sti->nb_index_entries++];
e->pos = current_offset;
e->timestamp = current_dts;
e->size = size;
@ -4092,8 +4101,8 @@ static void mov_build_index(MOVContext *mov, AVStream *st)
}
// Update start time of the stream.
if (st->start_time == AV_NOPTS_VALUE && st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && st->internal->nb_index_entries > 0) {
st->start_time = st->internal->index_entries[0].timestamp + sc->dts_shift;
if (st->start_time == AV_NOPTS_VALUE && st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && sti->nb_index_entries > 0) {
st->start_time = sti->index_entries[0].timestamp + sc->dts_shift;
if (sc->ctts_data) {
st->start_time += sc->ctts_data[0].duration;
}
@ -4328,7 +4337,7 @@ static int mov_read_trak(MOVContext *c, AVIOContext *pb, MOVAtom atom)
&& sc->stts_count > 3
&& sc->stts_count*10 > st->nb_frames
&& sc->time_scale == st->codecpar->sample_rate) {
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
}
/* Do not need those anymore. */
av_freep(&sc->chunk_offsets);
@ -4740,6 +4749,7 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
{
MOVFragment *frag = &c->fragment;
AVStream *st = NULL;
FFStream *sti = NULL;
MOVStreamContext *sc;
MOVStts *ctts_data;
uint64_t offset;
@ -4762,6 +4772,7 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
for (i = 0; i < c->fc->nb_streams; i++) {
if (c->fc->streams[i]->id == frag->track_id) {
st = c->fc->streams[i];
sti = ffstream(st);
break;
}
}
@ -4779,7 +4790,7 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
// A valid index_entry means the trun for the fragment was read
// and it's samples are in index_entries at the given position.
// New index entries will be inserted before the index_entry found.
index_entry_pos = st->internal->nb_index_entries;
index_entry_pos = sti->nb_index_entries;
for (i = c->frag_index.current + 1; i < c->frag_index.nb_items; i++) {
frag_stream_info = get_frag_stream_info(&c->frag_index, i, frag->track_id);
if (frag_stream_info && frag_stream_info->index_entry >= 0) {
@ -4788,7 +4799,7 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
break;
}
}
av_assert0(index_entry_pos <= st->internal->nb_index_entries);
av_assert0(index_entry_pos <= sti->nb_index_entries);
avio_r8(pb); /* version */
flags = avio_rb24(pb);
@ -4839,22 +4850,22 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
av_log(c->fc, AV_LOG_TRACE, "first sample flags 0x%x\n", first_sample_flags);
// realloc space for new index entries
if((uint64_t)st->internal->nb_index_entries + entries >= UINT_MAX / sizeof(AVIndexEntry)) {
entries = UINT_MAX / sizeof(AVIndexEntry) - st->internal->nb_index_entries;
if ((uint64_t)sti->nb_index_entries + entries >= UINT_MAX / sizeof(AVIndexEntry)) {
entries = UINT_MAX / sizeof(AVIndexEntry) - sti->nb_index_entries;
av_log(c->fc, AV_LOG_ERROR, "Failed to add index entry\n");
}
if (entries == 0)
return 0;
requested_size = (st->internal->nb_index_entries + entries) * sizeof(AVIndexEntry);
new_entries = av_fast_realloc(st->internal->index_entries,
&st->internal->index_entries_allocated_size,
requested_size = (sti->nb_index_entries + entries) * sizeof(AVIndexEntry);
new_entries = av_fast_realloc(sti->index_entries,
&sti->index_entries_allocated_size,
requested_size);
if (!new_entries)
return AVERROR(ENOMEM);
st->internal->index_entries= new_entries;
sti->index_entries= new_entries;
requested_size = (st->internal->nb_index_entries + entries) * sizeof(*sc->ctts_data);
requested_size = (sti->nb_index_entries + entries) * sizeof(*sc->ctts_data);
old_ctts_allocated_size = sc->ctts_allocated_size;
ctts_data = av_fast_realloc(sc->ctts_data, &sc->ctts_allocated_size,
requested_size);
@ -4868,12 +4879,12 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
memset((uint8_t*)(sc->ctts_data) + old_ctts_allocated_size, 0,
sc->ctts_allocated_size - old_ctts_allocated_size);
if (index_entry_pos < st->internal->nb_index_entries) {
if (index_entry_pos < sti->nb_index_entries) {
// Make hole in index_entries and ctts_data for new samples
memmove(st->internal->index_entries + index_entry_pos + entries,
st->internal->index_entries + index_entry_pos,
sizeof(*st->internal->index_entries) *
(st->internal->nb_index_entries - index_entry_pos));
memmove(sti->index_entries + index_entry_pos + entries,
sti->index_entries + index_entry_pos,
sizeof(*sti->index_entries) *
(sti->nb_index_entries - index_entry_pos));
memmove(sc->ctts_data + index_entry_pos + entries,
sc->ctts_data + index_entry_pos,
sizeof(*sc->ctts_data) * (sc->ctts_count - index_entry_pos));
@ -4882,15 +4893,15 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
}
}
st->internal->nb_index_entries += entries;
sc->ctts_count = st->internal->nb_index_entries;
sti->nb_index_entries += entries;
sc->ctts_count = sti->nb_index_entries;
// Record the index_entry position in frag_index of this fragment
if (frag_stream_info)
frag_stream_info->index_entry = index_entry_pos;
if (index_entry_pos > 0)
prev_dts = st->internal->index_entries[index_entry_pos-1].timestamp;
prev_dts = sti->index_entries[index_entry_pos-1].timestamp;
for (i = 0; i < entries && !pb->eof_reached; i++) {
unsigned sample_size = frag->size;
@ -4939,11 +4950,11 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
if (prev_dts >= dts)
index_entry_flags |= AVINDEX_DISCARD_FRAME;
st->internal->index_entries[index_entry_pos].pos = offset;
st->internal->index_entries[index_entry_pos].timestamp = dts;
st->internal->index_entries[index_entry_pos].size= sample_size;
st->internal->index_entries[index_entry_pos].min_distance= distance;
st->internal->index_entries[index_entry_pos].flags = index_entry_flags;
sti->index_entries[index_entry_pos].pos = offset;
sti->index_entries[index_entry_pos].timestamp = dts;
sti->index_entries[index_entry_pos].size = sample_size;
sti->index_entries[index_entry_pos].min_distance = distance;
sti->index_entries[index_entry_pos].flags = index_entry_flags;
sc->ctts_data[index_entry_pos].count = 1;
sc->ctts_data[index_entry_pos].duration = ctts_duration;
@ -4972,16 +4983,16 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
// EOF found before reading all entries. Fix the hole this would
// leave in index_entries and ctts_data
int gap = entries - i;
memmove(st->internal->index_entries + index_entry_pos,
st->internal->index_entries + index_entry_pos + gap,
sizeof(*st->internal->index_entries) *
(st->internal->nb_index_entries - (index_entry_pos + gap)));
memmove(sti->index_entries + index_entry_pos,
sti->index_entries + index_entry_pos + gap,
sizeof(*sti->index_entries) *
(sti->nb_index_entries - (index_entry_pos + gap)));
memmove(sc->ctts_data + index_entry_pos,
sc->ctts_data + index_entry_pos + gap,
sizeof(*sc->ctts_data) *
(sc->ctts_count - (index_entry_pos + gap)));
st->internal->nb_index_entries -= gap;
sti->nb_index_entries -= gap;
sc->ctts_count -= gap;
if (index_entry_pos < sc->current_sample) {
sc->current_sample -= gap;
@ -4994,11 +5005,11 @@ static int mov_read_trun(MOVContext *c, AVIOContext *pb, MOVAtom atom)
// fragment that overlap with AVINDEX_DISCARD_FRAME
prev_dts = AV_NOPTS_VALUE;
if (index_entry_pos > 0)
prev_dts = st->internal->index_entries[index_entry_pos-1].timestamp;
for (i = index_entry_pos; i < st->internal->nb_index_entries; i++) {
if (prev_dts < st->internal->index_entries[i].timestamp)
prev_dts = sti->index_entries[index_entry_pos-1].timestamp;
for (int i = index_entry_pos; i < sti->nb_index_entries; i++) {
if (prev_dts < sti->index_entries[i].timestamp)
break;
st->internal->index_entries[i].flags |= AVINDEX_DISCARD_FRAME;
sti->index_entries[i].flags |= AVINDEX_DISCARD_FRAME;
}
// If a hole was created to insert the new index_entries into,
@ -7156,15 +7167,15 @@ static int mov_probe(const AVProbeData *p)
static void mov_read_chapters(AVFormatContext *s)
{
MOVContext *mov = s->priv_data;
AVStream *st;
MOVStreamContext *sc;
int64_t cur_pos;
int i, j;
int chapter_track;
for (j = 0; j < mov->nb_chapter_tracks; j++) {
AVStream *st = NULL;
FFStream *sti = NULL;
chapter_track = mov->chapter_tracks[j];
st = NULL;
for (i = 0; i < s->nb_streams; i++)
if (s->streams[i]->id == chapter_track) {
st = s->streams[i];
@ -7174,15 +7185,16 @@ static void mov_read_chapters(AVFormatContext *s)
av_log(s, AV_LOG_ERROR, "Referenced QT chapter track not found\n");
continue;
}
sti = ffstream(st);
sc = st->priv_data;
cur_pos = avio_tell(sc->pb);
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
st->disposition |= AV_DISPOSITION_ATTACHED_PIC | AV_DISPOSITION_TIMED_THUMBNAILS;
if (st->internal->nb_index_entries) {
if (sti->nb_index_entries) {
// Retrieve the first frame, if possible
AVIndexEntry *sample = &st->internal->index_entries[0];
AVIndexEntry *sample = &sti->index_entries[0];
if (avio_seek(sc->pb, sample->pos, SEEK_SET) != sample->pos) {
av_log(s, AV_LOG_ERROR, "Failed to retrieve first frame\n");
goto finish;
@ -7195,9 +7207,9 @@ static void mov_read_chapters(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_DATA;
st->codecpar->codec_id = AV_CODEC_ID_BIN_DATA;
st->discard = AVDISCARD_ALL;
for (i = 0; i < st->internal->nb_index_entries; i++) {
AVIndexEntry *sample = &st->internal->index_entries[i];
int64_t end = i+1 < st->internal->nb_index_entries ? st->internal->index_entries[i+1].timestamp : st->duration;
for (int i = 0; i < sti->nb_index_entries; i++) {
AVIndexEntry *sample = &sti->index_entries[i];
int64_t end = i+1 < sti->nb_index_entries ? sti->index_entries[i+1].timestamp : st->duration;
uint8_t *title;
uint16_t ch;
int len, title_len;
@ -7266,14 +7278,15 @@ static int parse_timecode_in_framenum_format(AVFormatContext *s, AVStream *st,
static int mov_read_rtmd_track(AVFormatContext *s, AVStream *st)
{
MOVStreamContext *sc = st->priv_data;
FFStream *const sti = ffstream(st);
char buf[AV_TIMECODE_STR_SIZE];
int64_t cur_pos = avio_tell(sc->pb);
int hh, mm, ss, ff, drop;
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
return -1;
avio_seek(sc->pb, st->internal->index_entries->pos, SEEK_SET);
avio_seek(sc->pb, sti->index_entries->pos, SEEK_SET);
avio_skip(s->pb, 13);
hh = avio_r8(s->pb);
mm = avio_r8(s->pb);
@ -7291,14 +7304,15 @@ static int mov_read_rtmd_track(AVFormatContext *s, AVStream *st)
static int mov_read_timecode_track(AVFormatContext *s, AVStream *st)
{
MOVStreamContext *sc = st->priv_data;
FFStream *const sti = ffstream(st);
int flags = 0;
int64_t cur_pos = avio_tell(sc->pb);
uint32_t value;
if (!st->internal->nb_index_entries)
if (!sti->nb_index_entries)
return -1;
avio_seek(sc->pb, st->internal->index_entries->pos, SEEK_SET);
avio_seek(sc->pb, sti->index_entries->pos, SEEK_SET);
value = avio_rb32(s->pb);
if (sc->tmcd_flags & 0x0001) flags |= AV_TIMECODE_FLAG_DROPFRAME;
@ -7616,11 +7630,12 @@ static int mov_read_header(AVFormatContext *s)
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
MOVStreamContext *sc = st->priv_data;
fix_timescale(mov, sc);
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO &&
st->codecpar->codec_id == AV_CODEC_ID_AAC) {
st->internal->skip_samples = sc->start_pad;
sti->skip_samples = sc->start_pad;
}
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO && sc->nb_frames_for_fps > 0 && sc->duration_for_fps > 0)
av_reduce(&st->avg_frame_rate.num, &st->avg_frame_rate.den,
@ -7639,7 +7654,7 @@ static int mov_read_header(AVFormatContext *s)
mov->handbrake_version <= 1000000*0 + 1000*10 + 2 && // 0.10.2
st->codecpar->codec_id == AV_CODEC_ID_MP3) {
av_log(s, AV_LOG_VERBOSE, "Forcing full parsing for mp3 stream\n");
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
}
}
@ -7756,9 +7771,10 @@ static AVIndexEntry *mov_find_next_sample(AVFormatContext *s, AVStream **st)
int i;
for (i = 0; i < s->nb_streams; i++) {
AVStream *avst = s->streams[i];
FFStream *const avsti = ffstream(avst);
MOVStreamContext *msc = avst->priv_data;
if (msc->pb && msc->current_sample < avst->internal->nb_index_entries) {
AVIndexEntry *current_sample = &avst->internal->index_entries[msc->current_sample];
if (msc->pb && msc->current_sample < avsti->nb_index_entries) {
AVIndexEntry *current_sample = &avsti->index_entries[msc->current_sample];
int64_t dts = av_rescale(current_sample->timestamp, AV_TIME_BASE, msc->time_scale);
av_log(s, AV_LOG_TRACE, "stream %d, sample %d, dts %"PRId64"\n", i, msc->current_sample, dts);
if (!sample || (!(s->pb->seekable & AVIO_SEEKABLE_NORMAL) && current_sample->pos < sample->pos) ||
@ -7940,9 +7956,9 @@ static int mov_read_packet(AVFormatContext *s, AVPacket *pkt)
sc->has_palette = 0;
}
}
if (st->codecpar->codec_id == AV_CODEC_ID_MP3 && !st->internal->need_parsing && pkt->size > 4) {
if (st->codecpar->codec_id == AV_CODEC_ID_MP3 && !ffstream(st)->need_parsing && pkt->size > 4) {
if (ff_mpa_check_header(AV_RB32(pkt->data)) < 0)
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
}
}
@ -7961,8 +7977,8 @@ static int mov_read_packet(AVFormatContext *s, AVPacket *pkt)
sc->ctts_sample = 0;
}
} else {
int64_t next_dts = (sc->current_sample < st->internal->nb_index_entries) ?
st->internal->index_entries[sc->current_sample].timestamp : st->duration;
int64_t next_dts = (sc->current_sample < ffstream(st)->nb_index_entries) ?
ffstream(st)->index_entries[sc->current_sample].timestamp : st->duration;
if (next_dts >= pkt->dts)
pkt->duration = next_dts - pkt->dts;
@ -8030,6 +8046,7 @@ static int mov_seek_fragment(AVFormatContext *s, AVStream *st, int64_t timestamp
static int mov_seek_stream(AVFormatContext *s, AVStream *st, int64_t timestamp, int flags)
{
MOVStreamContext *sc = st->priv_data;
FFStream *const sti = ffstream(st);
int sample, time_sample, ret;
unsigned int i;
@ -8043,7 +8060,7 @@ static int mov_seek_stream(AVFormatContext *s, AVStream *st, int64_t timestamp,
sample = av_index_search_timestamp(st, timestamp, flags);
av_log(s, AV_LOG_TRACE, "stream %d, timestamp %"PRId64", sample %d\n", st->index, timestamp, sample);
if (sample < 0 && st->internal->nb_index_entries && timestamp < st->internal->index_entries[0].timestamp)
if (sample < 0 && sti->nb_index_entries && timestamp < sti->index_entries[0].timestamp)
sample = 0;
if (sample < 0) /* not sure what to do */
return AVERROR_INVALIDDATA;
@ -8084,8 +8101,9 @@ static int mov_seek_stream(AVFormatContext *s, AVStream *st, int64_t timestamp,
static int64_t mov_get_skip_samples(AVStream *st, int sample)
{
MOVStreamContext *sc = st->priv_data;
int64_t first_ts = st->internal->index_entries[0].timestamp;
int64_t ts = st->internal->index_entries[sample].timestamp;
FFStream *const sti = ffstream(st);
int64_t first_ts = sti->index_entries[0].timestamp;
int64_t ts = sti->index_entries[sample].timestamp;
int64_t off;
if (st->codecpar->codec_type != AVMEDIA_TYPE_AUDIO)
@ -8101,6 +8119,7 @@ static int mov_read_seek(AVFormatContext *s, int stream_index, int64_t sample_ti
{
MOVContext *mc = s->priv_data;
AVStream *st;
FFStream *sti;
int sample;
int i;
@ -8108,18 +8127,20 @@ static int mov_read_seek(AVFormatContext *s, int stream_index, int64_t sample_ti
return AVERROR_INVALIDDATA;
st = s->streams[stream_index];
sti = ffstream(st);
sample = mov_seek_stream(s, st, sample_time, flags);
if (sample < 0)
return sample;
if (mc->seek_individually) {
/* adjust seek timestamp to found sample timestamp */
int64_t seek_timestamp = st->internal->index_entries[sample].timestamp;
st->internal->skip_samples = mov_get_skip_samples(st, sample);
int64_t seek_timestamp = sti->index_entries[sample].timestamp;
sti->skip_samples = mov_get_skip_samples(st, sample);
for (i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
FFStream *const sti = ffstream(st);
int64_t timestamp;
st = s->streams[i];
if (stream_index == i)
continue;
@ -8127,7 +8148,7 @@ static int mov_read_seek(AVFormatContext *s, int stream_index, int64_t sample_ti
timestamp = av_rescale_q(seek_timestamp, s->streams[stream_index]->time_base, st->time_base);
sample = mov_seek_stream(s, st, timestamp, flags);
if (sample >= 0)
st->internal->skip_samples = mov_get_skip_samples(st, sample);
sti->skip_samples = mov_get_skip_samples(st, sample);
}
} else {
for (i = 0; i < s->nb_streams; i++) {

View File

@ -160,6 +160,7 @@ static void mp3_parse_info_tag(AVFormatContext *s, AVStream *st,
#define LAST_BITS(k, n) ((k) & ((1 << (n)) - 1))
#define MIDDLE_BITS(k, m, n) LAST_BITS((k) >> (m), ((n) - (m) + 1))
FFStream *const sti = ffstream(st);
uint16_t crc;
uint32_t v;
@ -256,13 +257,13 @@ static void mp3_parse_info_tag(AVFormatContext *s, AVStream *st,
mp3->start_pad = v>>12;
mp3-> end_pad = v&4095;
st->internal->start_skip_samples = mp3->start_pad + 528 + 1;
sti->start_skip_samples = mp3->start_pad + 528 + 1;
if (mp3->frames) {
st->internal->first_discard_sample = -mp3->end_pad + 528 + 1 + mp3->frames * (int64_t)spf;
st->internal->last_discard_sample = mp3->frames * (int64_t)spf;
sti->first_discard_sample = -mp3->end_pad + 528 + 1 + mp3->frames * (int64_t)spf;
sti->last_discard_sample = mp3->frames * (int64_t)spf;
}
if (!st->start_time)
st->start_time = av_rescale_q(st->internal->start_skip_samples,
st->start_time = av_rescale_q(sti->start_skip_samples,
(AVRational){1, c->sample_rate},
st->time_base);
av_log(s, AV_LOG_DEBUG, "pad %d %d\n", mp3->start_pad, mp3-> end_pad);
@ -363,6 +364,7 @@ static int mp3_read_header(AVFormatContext *s)
FFFormatContext *const si = ffformatcontext(s);
MP3DecContext *mp3 = s->priv_data;
AVStream *st;
FFStream *sti;
int64_t off;
int ret;
int i;
@ -373,10 +375,11 @@ static int mp3_read_header(AVFormatContext *s)
st = avformat_new_stream(s, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_MP3;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
st->start_time = 0;
// lcm of all mp3 sample rates
@ -434,8 +437,8 @@ static int mp3_read_header(AVFormatContext *s)
off = avio_tell(s->pb);
// the seek index is relative to the end of the xing vbr headers
for (i = 0; i < st->internal->nb_index_entries; i++)
st->internal->index_entries[i].pos += off;
for (int i = 0; i < sti->nb_index_entries; i++)
sti->index_entries[i].pos += off;
/* the parameters will be extracted from the compressed bitstream */
return 0;
@ -551,6 +554,7 @@ static int mp3_seek(AVFormatContext *s, int stream_index, int64_t timestamp,
MP3DecContext *mp3 = s->priv_data;
AVIndexEntry *ie, ie1;
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
int64_t best_pos;
int fast_seek = s->flags & AVFMT_FLAG_FAST_SEEK;
int64_t filesize = mp3->header_filesize;
@ -571,7 +575,7 @@ static int mp3_seek(AVFormatContext *s, int stream_index, int64_t timestamp,
if (ret < 0)
return ret;
ie = &st->internal->index_entries[ret];
ie = &sti->index_entries[ret];
} else if (fast_seek && st->duration > 0 && filesize > 0) {
if (!mp3->is_cbr)
av_log(s, AV_LOG_WARNING, "Using scaling to seek VBR MP3; may be imprecise.\n");

View File

@ -187,6 +187,7 @@ static int mpc_read_packet(AVFormatContext *s, AVPacket *pkt)
static int mpc_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int flags)
{
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
MPCContext *c = s->priv_data;
AVPacket pkt1, *pkt = &pkt1;
int ret;
@ -194,8 +195,8 @@ static int mpc_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
uint32_t lastframe;
/* if found, seek there */
if (index >= 0 && st->internal->index_entries[st->internal->nb_index_entries-1].timestamp >= timestamp - DELAY_FRAMES){
c->curframe = st->internal->index_entries[index].pos;
if (index >= 0 && sti->index_entries[sti->nb_index_entries-1].timestamp >= timestamp - DELAY_FRAMES) {
c->curframe = sti->index_entries[index].pos;
return 0;
}
/* if timestamp is out of bounds, return error */

View File

@ -324,12 +324,13 @@ static int mpc8_read_packet(AVFormatContext *s, AVPacket *pkt)
static int mpc8_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int flags)
{
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
int index = av_index_search_timestamp(st, timestamp, flags);
if(index < 0) return -1;
if (avio_seek(s->pb, st->internal->index_entries[index].pos, SEEK_SET) < 0)
if (avio_seek(s->pb, sti->index_entries[index].pos, SEEK_SET) < 0)
return -1;
avpriv_update_cur_dts(s, st, st->internal->index_entries[index].timestamp);
avpriv_update_cur_dts(s, st, sti->index_entries[index].timestamp);
return 0;
}

View File

@ -478,6 +478,7 @@ static int mpegps_read_packet(AVFormatContext *s,
{
MpegDemuxContext *m = s->priv_data;
AVStream *st;
FFStream *sti;
int len, startcode, i, es_type, ret;
int pcm_dvd = 0;
int request_probe= 0;
@ -614,6 +615,7 @@ skip:
st = avformat_new_stream(s, NULL);
if (!st)
goto skip;
sti = ffstream(st);
st->id = startcode;
st->codecpar->codec_type = type;
st->codecpar->codec_id = codec_id;
@ -623,8 +625,8 @@ skip:
st->codecpar->channel_layout = AV_CH_LAYOUT_MONO;
st->codecpar->sample_rate = 8000;
}
st->internal->request_probe = request_probe;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->request_probe = request_probe;
sti->need_parsing = AVSTREAM_PARSE_FULL;
found:
if (st->discard >= AVDISCARD_ALL)

View File

@ -875,15 +875,16 @@ static void mpegts_find_stream_type(AVStream *st,
uint32_t stream_type,
const StreamType *types)
{
FFStream *const sti = ffstream(st);
for (; types->stream_type; types++)
if (stream_type == types->stream_type) {
if (st->codecpar->codec_type != types->codec_type ||
st->codecpar->codec_id != types->codec_id) {
st->codecpar->codec_type = types->codec_type;
st->codecpar->codec_id = types->codec_id;
st->internal->need_context_update = 1;
sti->need_context_update = 1;
}
st->internal->request_probe = 0;
sti->request_probe = 0;
return;
}
}
@ -891,11 +892,12 @@ static void mpegts_find_stream_type(AVStream *st,
static int mpegts_set_stream_info(AVStream *st, PESContext *pes,
uint32_t stream_type, uint32_t prog_reg_desc)
{
FFStream *const sti = ffstream(st);
int old_codec_type = st->codecpar->codec_type;
int old_codec_id = st->codecpar->codec_id;
int old_codec_tag = st->codecpar->codec_tag;
if (avcodec_is_open(st->internal->avctx)) {
if (avcodec_is_open(sti->avctx)) {
av_log(pes->stream, AV_LOG_DEBUG, "cannot set stream info, internal codec is open\n");
return 0;
}
@ -904,7 +906,7 @@ static int mpegts_set_stream_info(AVStream *st, PESContext *pes,
st->priv_data = pes;
st->codecpar->codec_type = AVMEDIA_TYPE_DATA;
st->codecpar->codec_id = AV_CODEC_ID_NONE;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
pes->st = st;
pes->stream_type = stream_type;
@ -916,7 +918,7 @@ static int mpegts_set_stream_info(AVStream *st, PESContext *pes,
mpegts_find_stream_type(st, pes->stream_type, ISO_types);
if (pes->stream_type == 4 || pes->stream_type == 0x0f)
st->internal->request_probe = 50;
sti->request_probe = 50;
if ((prog_reg_desc == AV_RL32("HDMV") ||
prog_reg_desc == AV_RL32("HDPR")) &&
st->codecpar->codec_id == AV_CODEC_ID_NONE) {
@ -942,7 +944,7 @@ static int mpegts_set_stream_info(AVStream *st, PESContext *pes,
sub_st->priv_data = sub_pes;
sub_st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
sub_st->codecpar->codec_id = AV_CODEC_ID_AC3;
sub_st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(sub_st)->need_parsing = AVSTREAM_PARSE_FULL;
sub_pes->sub_st = pes->sub_st = sub_st;
}
}
@ -953,19 +955,19 @@ static int mpegts_set_stream_info(AVStream *st, PESContext *pes,
st->codecpar->codec_type = old_codec_type;
}
if ((st->codecpar->codec_id == AV_CODEC_ID_NONE ||
(st->internal->request_probe > 0 && st->internal->request_probe < AVPROBE_SCORE_STREAM_RETRY / 5)) &&
st->internal->probe_packets > 0 &&
(sti->request_probe > 0 && sti->request_probe < AVPROBE_SCORE_STREAM_RETRY / 5)) &&
sti->probe_packets > 0 &&
stream_type == STREAM_TYPE_PRIVATE_DATA) {
st->codecpar->codec_type = AVMEDIA_TYPE_DATA;
st->codecpar->codec_id = AV_CODEC_ID_BIN_DATA;
st->internal->request_probe = AVPROBE_SCORE_STREAM_RETRY / 5;
sti->request_probe = AVPROBE_SCORE_STREAM_RETRY / 5;
}
/* queue a context update if properties changed */
if (old_codec_type != st->codecpar->codec_type ||
old_codec_id != st->codecpar->codec_id ||
old_codec_tag != st->codecpar->codec_tag)
st->internal->need_context_update = 1;
sti->need_context_update = 1;
return 0;
}
@ -1198,13 +1200,14 @@ static int mpegts_push_data(MpegTSFilter *filter,
code != 0x1f0 && code != 0x1f1 && /* ECM, EMM */
code != 0x1ff && code != 0x1f2 && /* program_stream_directory, DSMCC_stream */
code != 0x1f8) { /* ITU-T Rec. H.222.1 type E stream */
FFStream *const pes_sti = ffstream(pes->st);
pes->state = MPEGTS_PESHEADER;
if (pes->st->codecpar->codec_id == AV_CODEC_ID_NONE && !pes->st->internal->request_probe) {
if (pes->st->codecpar->codec_id == AV_CODEC_ID_NONE && !pes_sti->request_probe) {
av_log(pes->stream, AV_LOG_TRACE,
"pid=%x stream_type=%x probing\n",
pes->pid,
pes->stream_type);
pes->st->internal->request_probe = 1;
pes_sti->request_probe = 1;
}
} else {
pes->pes_header_size = 6;
@ -1326,8 +1329,11 @@ skip:
int64_t pcr = f->last_pcr / 300;
pcr_found = 1;
if (st) {
pes->st->internal->pts_wrap_reference = st->internal->pts_wrap_reference;
pes->st->internal->pts_wrap_behavior = st->internal->pts_wrap_behavior;
const FFStream *const sti = ffstream(st);
FFStream *const pes_sti = ffstream(pes->st);
pes_sti->pts_wrap_reference = sti->pts_wrap_reference;
pes_sti->pts_wrap_behavior = sti->pts_wrap_behavior;
}
if (pes->dts == AV_NOPTS_VALUE || pes->dts < pcr) {
pes->pts = pes->dts = pcr;
@ -1699,6 +1705,7 @@ static void m4sl_cb(MpegTSFilter *filter, const uint8_t *section,
for (i = 0; i < mp4_descr_count; i++) {
PESContext *pes;
AVStream *st;
FFStream *sti;
FFIOContext pb;
if (ts->pids[pid]->es_id != mp4_descr[i].es_id)
continue;
@ -1710,6 +1717,7 @@ static void m4sl_cb(MpegTSFilter *filter, const uint8_t *section,
st = pes->st;
if (!st)
continue;
sti = ffstream(st);
pes->sl = mp4_descr[i].sl;
@ -1719,13 +1727,13 @@ static void m4sl_cb(MpegTSFilter *filter, const uint8_t *section,
ff_mp4_read_dec_config_descr(s, st, &pb.pub);
if (st->codecpar->codec_id == AV_CODEC_ID_AAC &&
st->codecpar->extradata_size > 0)
st->internal->need_parsing = 0;
sti->need_parsing = 0;
if (st->codecpar->codec_id == AV_CODEC_ID_H264 &&
st->codecpar->extradata_size > 0)
st->internal->need_parsing = 0;
sti->need_parsing = 0;
st->codecpar->codec_type = avcodec_get_type(st->codecpar->codec_id);
st->internal->need_context_update = 1;
sti->need_context_update = 1;
}
}
for (i = 0; i < mp4_descr_count; i++)
@ -1785,6 +1793,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
Mp4Descr *mp4_descr, int mp4_descr_count, int pid,
MpegTSContext *ts)
{
FFStream *const sti = ffstream(st);
const uint8_t *desc_end;
int desc_len, desc_tag, desc_es_id, ext_desc_tag, channels, channel_config_code;
char language[252];
@ -1802,7 +1811,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
av_log(fc, AV_LOG_TRACE, "tag: 0x%02x len=%d\n", desc_tag, desc_len);
if ((st->codecpar->codec_id == AV_CODEC_ID_NONE || st->internal->request_probe > 0) &&
if ((st->codecpar->codec_id == AV_CODEC_ID_NONE || sti->request_probe > 0) &&
stream_type == STREAM_TYPE_PRIVATE_DATA)
mpegts_find_stream_type(st, desc_tag, DESC_types);
@ -1828,8 +1837,8 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
ff_mp4_read_dec_config_descr(fc, st, &pb.pub);
if (st->codecpar->codec_id == AV_CODEC_ID_AAC &&
st->codecpar->extradata_size > 0) {
st->internal->need_parsing = 0;
st->internal->need_context_update = 1;
sti->need_parsing = 0;
sti->need_context_update = 1;
}
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG4SYSTEMS)
mpegts_open_section_filter(ts, pid, m4sl_cb, ts, 1);
@ -1840,8 +1849,8 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
break;
if (mp4_descr_count > 0 &&
(st->codecpar->codec_id == AV_CODEC_ID_AAC_LATM ||
(st->internal->request_probe == 0 && st->codecpar->codec_id == AV_CODEC_ID_NONE) ||
st->internal->request_probe > 0) &&
(sti->request_probe == 0 && st->codecpar->codec_id == AV_CODEC_ID_NONE) ||
sti->request_probe > 0) &&
mp4_descr->dec_config_descr_len && mp4_descr->es_id == pid) {
FFIOContext pb;
ffio_init_context(&pb, mp4_descr->dec_config_descr,
@ -1850,9 +1859,9 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
ff_mp4_read_dec_config_descr(fc, st, &pb.pub);
if (st->codecpar->codec_id == AV_CODEC_ID_AAC &&
st->codecpar->extradata_size > 0) {
st->internal->request_probe = st->internal->need_parsing = 0;
sti->request_probe = sti->need_parsing = 0;
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->internal->need_context_update = 1;
sti->need_context_update = 1;
}
}
break;
@ -1893,7 +1902,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
language[i * 4 - 1] = 0;
av_dict_set(&st->metadata, "language", language, 0);
st->internal->need_context_update = 1;
sti->need_context_update = 1;
}
}
break;
@ -1957,7 +1966,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
language[i * 4 - 1] = 0;
av_dict_set(&st->metadata, "language", language, 0);
st->internal->need_context_update = 1;
sti->need_context_update = 1;
}
}
break;
@ -1990,14 +1999,14 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
case REGISTRATION_DESCRIPTOR:
st->codecpar->codec_tag = bytestream_get_le32(pp);
av_log(fc, AV_LOG_TRACE, "reg_desc=%.4s\n", (char *)&st->codecpar->codec_tag);
if (st->codecpar->codec_id == AV_CODEC_ID_NONE || st->internal->request_probe > 0) {
if (st->codecpar->codec_id == AV_CODEC_ID_NONE || sti->request_probe > 0) {
mpegts_find_stream_type(st, st->codecpar->codec_tag, REGD_types);
if (st->codecpar->codec_tag == MKTAG('B', 'S', 'S', 'D'))
st->internal->request_probe = 50;
sti->request_probe = 50;
}
break;
case 0x52: /* stream identifier descriptor */
st->internal->stream_identifier = 1 + get8(pp, desc_end);
sti->stream_identifier = 1 + get8(pp, desc_end);
break;
case METADATA_DESCRIPTOR:
if (get16(pp, desc_end) == 0xFFFF)
@ -2037,8 +2046,8 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
} else {
avpriv_request_sample(fc, "Opus in MPEG-TS - channel_config_code > 0x8");
}
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
st->internal->need_context_update = 1;
sti->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_context_update = 1;
}
}
if (ext_desc_tag == 0x06) { /* supplementary audio descriptor */
@ -2116,7 +2125,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
// Listing of data_component_ids is in STD-B10, part 2, Annex J.
// Component tag limits are documented in TR-B14, fascicle 2,
// Vol. 3, Section 2, 4.2.8.1
int actual_component_tag = st->internal->stream_identifier - 1;
int actual_component_tag = sti->stream_identifier - 1;
int picked_profile = FF_PROFILE_UNKNOWN;
int data_component_id = get16(pp, desc_end);
if (data_component_id < 0)
@ -2148,7 +2157,7 @@ int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type
st->codecpar->codec_type = AVMEDIA_TYPE_SUBTITLE;
st->codecpar->codec_id = AV_CODEC_ID_ARIB_CAPTION;
st->codecpar->profile = picked_profile;
st->internal->request_probe = 0;
sti->request_probe = 0;
}
break;
case 0xb0: /* DOVI video stream descriptor */

View File

@ -82,7 +82,7 @@ static int msf_read_header(AVFormatContext *s)
AV_WL16(st->codecpar->extradata+8, codec == 4 ? 1 : 0); /* joint stereo (repeat?) */
AV_WL16(st->codecpar->extradata+10, 1);
st->codecpar->codec_id = AV_CODEC_ID_ATRAC3; break;
case 7: st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
case 7: ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
st->codecpar->codec_id = AV_CODEC_ID_MP3; break;
default:
avpriv_request_sample(s, "Codec %d", codec);

View File

@ -185,7 +185,7 @@ static int mtv_read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_MP3;
st->codecpar->bit_rate = mtv->audio_br;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
// Jump over header

View File

@ -253,7 +253,7 @@ static int init_muxer(AVFormatContext *s, AVDictionary **options)
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVCodecParameters *const par = st->codecpar;
const AVCodecDescriptor *desc;
@ -386,7 +386,7 @@ static int init_pts(AVFormatContext *s)
/* init PTS generation */
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int64_t den = AV_NOPTS_VALUE;
switch (st->codecpar->codec_type) {
@ -509,7 +509,7 @@ FF_DISABLE_DEPRECATION_WARNINGS
static int compute_muxer_pkt_fields(AVFormatContext *s, AVStream *st, AVPacket *pkt)
{
FFFormatContext *const si = ffformatcontext(s);
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int delay = st->codecpar->video_delay;
int frame_size;
@ -645,7 +645,7 @@ static int write_packet(AVFormatContext *s, AVPacket *pkt)
{
FFFormatContext *const si = ffformatcontext(s);
AVStream *const st = s->streams[pkt->stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int ret;
// If the timestamp offsetting below is adjusted, adjust
@ -743,7 +743,7 @@ static int check_packet(AVFormatContext *s, AVPacket *pkt)
static int prepare_input_packet(AVFormatContext *s, AVStream *st, AVPacket *pkt)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
#if !FF_API_COMPUTE_PKT_FIELDS2
/* sanitize the timestamps */
if (!(s->oformat->flags & AVFMT_NOTIMESTAMPS)) {
@ -799,7 +799,7 @@ int ff_interleave_add_packet(AVFormatContext *s, AVPacket *pkt,
FFFormatContext *const si = ffformatcontext(s);
PacketList **next_point, *this_pktl;
AVStream *st = s->streams[pkt->stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int chunked = s->max_chunk_size || s->max_chunk_duration;
this_pktl = av_malloc(sizeof(PacketList));
@ -912,7 +912,7 @@ int ff_interleave_packet_per_dts(AVFormatContext *s, AVPacket *out,
for (unsigned i = 0; i < s->nb_streams; i++) {
const AVStream *const st = s->streams[i];
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = cffstream(st);
const AVCodecParameters *const par = st->codecpar;
if (sti->last_in_packet_buffer) {
++stream_count;
@ -939,7 +939,7 @@ int ff_interleave_packet_per_dts(AVFormatContext *s, AVPacket *out,
for (unsigned i = 0; i < s->nb_streams; i++) {
const AVStream *const st = s->streams[i];
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = cffstream(st);
const PacketList *last = sti->last_in_packet_buffer;
int64_t last_dts;
@ -977,7 +977,7 @@ int ff_interleave_packet_per_dts(AVFormatContext *s, AVPacket *out,
PacketList *pktl = si->packet_buffer;
AVPacket *const top_pkt = &pktl->pkt;
AVStream *const st = s->streams[top_pkt->stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int64_t top_dts = av_rescale_q(top_pkt->dts, st->time_base,
AV_TIME_BASE_Q);
@ -1000,7 +1000,7 @@ int ff_interleave_packet_per_dts(AVFormatContext *s, AVPacket *out,
if (stream_count && flush) {
PacketList *pktl = si->packet_buffer;
AVStream *const st = s->streams[pktl->pkt.stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
*out = pktl->pkt;
@ -1026,7 +1026,7 @@ int ff_get_muxer_ts_offset(AVFormatContext *s, int stream_index, int64_t *offset
return AVERROR(EINVAL);
st = s->streams[stream_index];
*offset = st->internal->mux_ts_offset;
*offset = ffstream(st)->mux_ts_offset;
if (s->output_ts_offset)
*offset += av_rescale_q(s->output_ts_offset, AV_TIME_BASE_Q, st->time_base);
@ -1064,7 +1064,7 @@ static int interleave_packet(AVFormatContext *s, AVPacket *out, AVPacket *in, in
return ff_interleave_packet_per_dts(s, out, in, flush);
}
static int check_bitstream(AVFormatContext *s, AVStreamInternal *sti, AVPacket *pkt)
static int check_bitstream(AVFormatContext *s, FFStream *sti, AVPacket *pkt)
{
int ret;
@ -1128,7 +1128,7 @@ static int write_packet_common(AVFormatContext *s, AVStream *st, AVPacket *pkt,
static int write_packets_from_bsfs(AVFormatContext *s, AVStream *st, AVPacket *pkt, int interleaved)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVBSFContext *const bsfc = sti->bsfc;
int ret;
@ -1162,12 +1162,12 @@ static int write_packets_from_bsfs(AVFormatContext *s, AVStream *st, AVPacket *p
static int write_packets_common(AVFormatContext *s, AVPacket *pkt, int interleaved)
{
AVStream *st;
AVStreamInternal *sti;
FFStream *sti;
int ret = check_packet(s, pkt);
if (ret < 0)
return ret;
st = s->streams[pkt->stream_index];
sti = st->internal;
sti = ffstream(st);
ret = prepare_input_packet(s, st, pkt);
if (ret < 0)
@ -1256,7 +1256,7 @@ int av_write_trailer(AVFormatContext *s)
av_packet_unref(pkt);
for (unsigned i = 0; i < s->nb_streams; i++) {
if (s->streams[i]->internal->bsfc) {
if (ffstream(s->streams[i])->bsfc) {
ret1 = write_packets_from_bsfs(s, s->streams[i], pkt, 1/*interleaved*/);
if (ret1 < 0)
av_packet_unref(pkt);
@ -1286,7 +1286,7 @@ int av_write_trailer(AVFormatContext *s)
ret = s->pb ? s->pb->error : 0;
for (unsigned i = 0; i < s->nb_streams; i++) {
av_freep(&s->streams[i]->priv_data);
av_freep(&s->streams[i]->internal->index_entries);
av_freep(&ffstream(s->streams[i])->index_entries);
}
if (s->oformat->priv_class)
av_opt_free(s->priv_data);

View File

@ -436,13 +436,14 @@ static int mv_read_packet(AVFormatContext *avctx, AVPacket *pkt)
MvContext *mv = avctx->priv_data;
AVIOContext *pb = avctx->pb;
AVStream *st = avctx->streams[mv->stream_index];
FFStream *const sti = ffstream(st);
const AVIndexEntry *index;
int frame = mv->frame[mv->stream_index];
int64_t ret;
uint64_t pos;
if (frame < st->internal->nb_index_entries) {
index = &st->internal->index_entries[frame];
if (frame < sti->nb_index_entries) {
index = &sti->index_entries[frame];
pos = avio_tell(pb);
if (index->pos > pos)
avio_skip(pb, index->pos - pos);

View File

@ -1793,7 +1793,7 @@ static int mxf_compute_ptses_fake_index(MXFContext *mxf, MXFIndexTable *index_ta
* 6: 5 5
*
* We do this by bucket sorting x by x+TemporalOffset[x] into mxf->ptses,
* then settings mxf->internal->first_dts = -max(TemporalOffset[x]).
* then settings ffstream(mxf)->first_dts = -max(TemporalOffset[x]).
* The latter makes DTS <= PTS.
*/
for (i = x = 0; i < index_table->nb_segments; i++) {
@ -2328,6 +2328,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
const MXFCodecUL *container_ul = NULL;
const MXFCodecUL *pix_fmt_ul = NULL;
AVStream *st;
FFStream *sti;
AVTimecode tc;
int flags;
@ -2435,6 +2436,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
ret = AVERROR(ENOMEM);
goto fail_and_free;
}
sti = ffstream(st);
st->id = material_track->track_id;
st->priv_data = source_track;
@ -2620,7 +2622,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
}
}
}
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
if (material_track->sequence->origin) {
av_dict_set_int(&st->metadata, "material_track_origin", material_track->sequence->origin, 0);
}
@ -2628,7 +2630,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
av_dict_set_int(&st->metadata, "source_track_origin", source_track->sequence->origin, 0);
}
if (descriptor->aspect_ratio.num && descriptor->aspect_ratio.den)
st->internal->display_aspect_ratio = descriptor->aspect_ratio;
sti->display_aspect_ratio = descriptor->aspect_ratio;
st->codecpar->color_range = mxf_get_color_range(mxf, descriptor);
st->codecpar->color_primaries = mxf_get_codec_ul(ff_mxf_color_primaries_uls, &descriptor->color_primaries_ul)->id;
st->codecpar->color_trc = mxf_get_codec_ul(ff_mxf_color_trc_uls, &descriptor->color_trc_ul)->id;
@ -2685,7 +2687,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
else if (descriptor->bits_per_sample == 32)
st->codecpar->codec_id = AV_CODEC_ID_PCM_S32BE;
} else if (st->codecpar->codec_id == AV_CODEC_ID_MP2) {
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
}
st->codecpar->bits_per_coded_sample = av_get_bits_per_sample(st->codecpar->codec_id);
} else if (st->codecpar->codec_type == AVMEDIA_TYPE_DATA) {
@ -2719,7 +2721,7 @@ static int mxf_parse_structural_metadata(MXFContext *mxf)
}
if (st->codecpar->codec_type != AVMEDIA_TYPE_DATA && source_track->wrapping != FrameWrapped) {
/* TODO: decode timestamps */
st->internal->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
sti->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
}
}
@ -3680,7 +3682,7 @@ static int mxf_read_packet(AVFormatContext *s, AVPacket *pkt)
if (next_ofs <= 0) {
// If we have no way to packetize the data, then return it in chunks...
if (klv.next_klv - klv.length == pos && max_data_size > MXF_MAX_CHUNK_SIZE) {
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_request_sample(s, "Huge KLV without proper index in non-frame wrapped essence");
}
size = FFMIN(max_data_size, MXF_MAX_CHUNK_SIZE);

View File

@ -3104,7 +3104,7 @@ static int mxf_interleave_get_packet(AVFormatContext *s, AVPacket *out, AVPacket
int i, stream_count = 0;
for (i = 0; i < s->nb_streams; i++)
stream_count += !!s->streams[i]->internal->last_in_packet_buffer;
stream_count += !!ffstream(s->streams[i])->last_in_packet_buffer;
if (stream_count && (s->nb_streams == stream_count || flush)) {
PacketList *pktl = si->packet_buffer;
@ -3115,8 +3115,8 @@ static int mxf_interleave_get_packet(AVFormatContext *s, AVPacket *out, AVPacket
if (!stream_count || pktl->pkt.stream_index == 0)
break;
// update last packet in packet buffer
if (s->streams[pktl->pkt.stream_index]->internal->last_in_packet_buffer != pktl)
s->streams[pktl->pkt.stream_index]->internal->last_in_packet_buffer = pktl;
if (ffstream(s->streams[pktl->pkt.stream_index])->last_in_packet_buffer != pktl)
ffstream(s->streams[pktl->pkt.stream_index])->last_in_packet_buffer = pktl;
last = pktl;
pktl = pktl->next;
stream_count--;
@ -3141,8 +3141,8 @@ static int mxf_interleave_get_packet(AVFormatContext *s, AVPacket *out, AVPacket
*out = pktl->pkt;
av_log(s, AV_LOG_TRACE, "out st:%d dts:%"PRId64"\n", (*out).stream_index, (*out).dts);
si->packet_buffer = pktl->next;
if(s->streams[pktl->pkt.stream_index]->internal->last_in_packet_buffer == pktl)
s->streams[pktl->pkt.stream_index]->internal->last_in_packet_buffer= NULL;
if (ffstream(s->streams[pktl->pkt.stream_index])->last_in_packet_buffer == pktl)
ffstream(s->streams[pktl->pkt.stream_index])->last_in_packet_buffer = NULL;
if (!si->packet_buffer)
si->packet_buffer_end = NULL;
av_freep(&pktl);

View File

@ -53,7 +53,7 @@ static int nc_read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG4;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_set_pts_info(st, 64, 1, 100);

View File

@ -461,7 +461,7 @@ static int nsv_parse_NSVs_header(AVFormatContext *s)
st->codecpar->codec_tag = atag;
st->codecpar->codec_id = ff_codec_get_id(nsv_codec_audio_tags, atag);
st->internal->need_parsing = AVSTREAM_PARSE_FULL; /* for PCM we will read a chunk later and put correct info */
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL; /* for PCM we will read a chunk later and put correct info */
/* set timebase to common denominator of ms and framerate */
avpriv_set_pts_info(st, 64, 1, framerate.num*1000);
@ -609,7 +609,7 @@ null_chunk_retry:
asize-=4;
av_log(s, AV_LOG_TRACE, "NSV RAWAUDIO: bps %d, nchan %d, srate %d\n", bps, channels, samplerate);
if (fill_header) {
st[NSV_ST_AUDIO]->internal->need_parsing = AVSTREAM_PARSE_NONE; /* we know everything */
ffstream(st[NSV_ST_AUDIO])->need_parsing = AVSTREAM_PARSE_NONE; /* we know everything */
if (bps != 16) {
av_log(s, AV_LOG_TRACE, "NSV AUDIO bit/sample != 16 (%d)!!!\n", bps);
}
@ -669,6 +669,7 @@ static int nsv_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
{
NSVContext *nsv = s->priv_data;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
NSVStream *nst = st->priv_data;
int index;
@ -676,10 +677,10 @@ static int nsv_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
if(index < 0)
return -1;
if (avio_seek(s->pb, st->internal->index_entries[index].pos, SEEK_SET) < 0)
if (avio_seek(s->pb, sti->index_entries[index].pos, SEEK_SET) < 0)
return -1;
nst->frame_offset = st->internal->index_entries[index].timestamp;
nst->frame_offset = sti->index_entries[index].timestamp;
nsv->state = NSV_UNSYNC;
return 0;
}

View File

@ -1086,7 +1086,7 @@ static int decode_frame(NUTContext *nut, AVPacket *pkt, int frame_code)
stc->skip_until_key_frame = 0;
discard = s->streams[stream_id]->discard;
last_IP_pts = s->streams[stream_id]->internal->last_IP_pts;
last_IP_pts = ffstream(s->streams[stream_id])->last_IP_pts;
if ((discard >= AVDISCARD_NONKEY && !(stc->last_flags & FLAG_KEY)) ||
(discard >= AVDISCARD_BIDIR && last_IP_pts != AV_NOPTS_VALUE &&
last_IP_pts > pts) ||
@ -1225,6 +1225,7 @@ static int read_seek(AVFormatContext *s, int stream_index,
{
NUTContext *nut = s->priv_data;
AVStream *st = s->streams[stream_index];
FFStream *const sti = ffstream(st);
Syncpoint dummy = { .ts = pts * av_q2d(st->time_base) * AV_TIME_BASE };
Syncpoint nopts_sp = { .ts = AV_NOPTS_VALUE, .back_ptr = AV_NOPTS_VALUE };
Syncpoint *sp, *next_node[2] = { &nopts_sp, &nopts_sp };
@ -1235,15 +1236,15 @@ static int read_seek(AVFormatContext *s, int stream_index,
return AVERROR(ENOSYS);
}
if (st->internal->index_entries) {
if (sti->index_entries) {
int index = av_index_search_timestamp(st, pts, flags);
if (index < 0)
index = av_index_search_timestamp(st, pts, flags ^ AVSEEK_FLAG_BACKWARD);
if (index < 0)
return -1;
pos2 = st->internal->index_entries[index].pos;
ts = st->internal->index_entries[index].timestamp;
pos2 = sti->index_entries[index].pos;
ts = sti->index_entries[index].timestamp;
} else {
av_tree_find(nut->syncpoints, &dummy, ff_nut_sp_pts_cmp,
(void **) next_node);

View File

@ -1006,6 +1006,7 @@ static int nut_write_packet(AVFormatContext *s, AVPacket *pkt)
ff_nut_reset_ts(nut, *nus->time_base, pkt->dts);
for (i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
int64_t dts_tb = av_rescale_rnd(pkt->dts,
nus->time_base->num * (int64_t)nut->stream[i].time_base->den,
nus->time_base->den * (int64_t)nut->stream[i].time_base->num,
@ -1013,12 +1014,12 @@ static int nut_write_packet(AVFormatContext *s, AVPacket *pkt)
int index = av_index_search_timestamp(st, dts_tb,
AVSEEK_FLAG_BACKWARD);
if (index >= 0) {
sp_pos = FFMIN(sp_pos, st->internal->index_entries[index].pos);
if (!nut->write_index && 2*index > st->internal->nb_index_entries) {
memmove(st->internal->index_entries,
st->internal->index_entries + index,
sizeof(*st->internal->index_entries) * (st->internal->nb_index_entries - index));
st->internal->nb_index_entries -= index;
sp_pos = FFMIN(sp_pos, sti->index_entries[index].pos);
if (!nut->write_index && 2*index > sti->nb_index_entries) {
memmove(sti->index_entries,
sti->index_entries + index,
sizeof(*sti->index_entries) * (sti->nb_index_entries - index));
sti->nb_index_entries -= index;
}
}
}

View File

@ -133,7 +133,7 @@ static int get_codec_data(AVFormatContext *s, AVIOContext *pb, AVStream *vst,
}
ast->codecpar->codec_id = id;
ast->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(ast)->need_parsing = AVSTREAM_PARSE_FULL;
} else
avio_skip(pb, 4 * 4);

View File

@ -59,7 +59,7 @@ flac_header (AVFormatContext * s, int idx)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_FLAC;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
if ((ret = ff_alloc_extradata(st->codecpar, FLAC_STREAMINFO_SIZE)) < 0)
return ret;

View File

@ -39,6 +39,7 @@ ogm_header(AVFormatContext *s, int idx)
struct ogg *ogg = s->priv_data;
struct ogg_stream *os = ogg->streams + idx;
AVStream *st = s->streams[idx];
FFStream *const sti = ffstream(st);
GetByteContext p;
uint64_t time_unit;
uint64_t spu;
@ -60,7 +61,7 @@ ogm_header(AVFormatContext *s, int idx)
st->codecpar->codec_id = ff_codec_get_id(ff_codec_bmp_tags, tag);
st->codecpar->codec_tag = tag;
if (st->codecpar->codec_id == AV_CODEC_ID_MPEG4)
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
} else if (bytestream2_peek_byte(&p) == 't') {
st->codecpar->codec_type = AVMEDIA_TYPE_SUBTITLE;
st->codecpar->codec_id = AV_CODEC_ID_TEXT;
@ -76,7 +77,7 @@ ogm_header(AVFormatContext *s, int idx)
st->codecpar->codec_id = ff_codec_get_id(ff_codec_wav_tags, cid);
// our parser completely breaks AAC in Ogg
if (st->codecpar->codec_id != AV_CODEC_ID_AAC)
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
}
size = bytestream2_get_le32(&p);
@ -116,7 +117,7 @@ ogm_header(AVFormatContext *s, int idx)
}
// Update internal avctx with changes to codecpar above.
st->internal->need_context_update = 1;
sti->need_context_update = 1;
} else if (bytestream2_peek_byte(&p) == 3) {
bytestream2_skip(&p, 7);
if (bytestream2_get_bytes_left(&p) > 1)

View File

@ -112,7 +112,7 @@ static int theora_header(AVFormatContext *s, int idx)
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_THEORA;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
}
break;
case 0x81:

View File

@ -61,7 +61,7 @@ static int vp8_header(AVFormatContext *s, int idx)
avpriv_set_pts_info(st, 64, framerate.den, framerate.num);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_VP8;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
break;
case 0x02:
if (p[6] != 0x20)

View File

@ -515,7 +515,7 @@ static int oma_read_header(AVFormatContext *s)
avpriv_set_pts_info(st, 64, 1, samplerate);
break;
case OMA_CODECID_MP3:
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
framesize = 1024;
break;
case OMA_CODECID_LPCM:

View File

@ -79,7 +79,7 @@ int ff_pcm_read_seek(AVFormatContext *s,
pos *= block_align;
/* recompute exact position */
st->internal->cur_dts = av_rescale(pos, st->time_base.den, byte_rate * (int64_t)st->time_base.num);
ffstream(st)->cur_dts = av_rescale(pos, st->time_base.den, byte_rate * (int64_t)st->time_base.num);
if ((ret = avio_seek(s->pb, pos + ffformatcontext(s)->data_offset, SEEK_SET)) < 0)
return ret;
return 0;

View File

@ -62,7 +62,7 @@ static int pva_read_header(AVFormatContext *s) {
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG2VIDEO;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_set_pts_info(st, 32, 1, 90000);
av_add_index_entry(st, 0, 0, 0, 0, AVINDEX_KEYFRAME);
@ -70,7 +70,7 @@ static int pva_read_header(AVFormatContext *s) {
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_MP2;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_set_pts_info(st, 33, 1, 90000);
av_add_index_entry(st, 0, 0, 0, 0, AVINDEX_KEYFRAME);

View File

@ -59,7 +59,7 @@ int ff_raw_audio_read_header(AVFormatContext *s)
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = s->iformat->raw_codec_id;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
st->start_time = 0;
/* the parameters will be extracted from the compressed bitstream */
@ -70,6 +70,7 @@ int ff_raw_audio_read_header(AVFormatContext *s)
int ff_raw_video_read_header(AVFormatContext *s)
{
AVStream *st;
FFStream *sti;
FFRawVideoDemuxerContext *s1 = s->priv_data;
int ret = 0;
@ -79,12 +80,13 @@ int ff_raw_video_read_header(AVFormatContext *s)
ret = AVERROR(ENOMEM);
goto fail;
}
sti = ffstream(st);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = s->iformat->raw_codec_id;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
st->internal->avctx->framerate = s1->framerate;
sti->avctx->framerate = s1->framerate;
avpriv_set_pts_info(st, 64, 1, 1200000);
fail:

View File

@ -416,7 +416,7 @@ rdt_parse_sdp_line (AVFormatContext *s, int st_index,
if (av_strstart(p, "OpaqueData:buffer;", &p)) {
rdt->mlti_data = rdt_parse_b64buf(&rdt->mlti_data_size, p);
} else if (av_strstart(p, "StartTime:integer;", &p))
stream->internal->first_dts = atoi(p);
ffstream(stream)->first_dts = atoi(p);
else if (av_strstart(p, "ASMRuleBook:string;", &p)) {
int n, first = -1;
@ -466,7 +466,7 @@ add_dstream(AVFormatContext *s, AVStream *orig_st)
return NULL;
st->id = orig_st->id;
st->codecpar->codec_type = orig_st->codecpar->codec_type;
st->internal->first_dts = orig_st->internal->first_dts;
ffstream(st)->first_dts = ffstream(orig_st)->first_dts;
return st;
}

View File

@ -237,9 +237,10 @@ static int rl2_read_packet(AVFormatContext *s,
/** check if there is a valid video or audio entry that can be used */
for(i=0; i<s->nb_streams; i++){
if(rl2->index_pos[i] < s->streams[i]->internal->nb_index_entries
&& s->streams[i]->internal->index_entries[ rl2->index_pos[i] ].pos < pos){
sample = &s->streams[i]->internal->index_entries[ rl2->index_pos[i] ];
const FFStream *const sti = ffstream(s->streams[i]);
if (rl2->index_pos[i] < sti->nb_index_entries
&& sti->index_entries[ rl2->index_pos[i] ].pos < pos) {
sample = &sti->index_entries[ rl2->index_pos[i] ];
pos= sample->pos;
stream_id= i;
}
@ -283,7 +284,7 @@ static int rl2_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
return -1;
rl2->index_pos[stream_index] = index;
timestamp = st->internal->index_entries[index].timestamp;
timestamp = ffstream(st)->index_entries[index].timestamp;
for(i=0; i < s->nb_streams; i++){
AVStream *st2 = s->streams[i];

View File

@ -122,6 +122,7 @@ void ff_rm_free_rmstream (RMStream *rms)
static int rm_read_audio_stream_info(AVFormatContext *s, AVIOContext *pb,
AVStream *st, RMStream *ast, int read_all)
{
FFStream *const sti = ffstream(st);
char buf[256];
uint32_t version;
int ret;
@ -202,7 +203,7 @@ static int rm_read_audio_stream_info(AVFormatContext *s, AVIOContext *pb,
switch (st->codecpar->codec_id) {
case AV_CODEC_ID_AC3:
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
sti->need_parsing = AVSTREAM_PARSE_FULL;
break;
case AV_CODEC_ID_RA_288:
st->codecpar->extradata_size= 0;
@ -211,7 +212,7 @@ static int rm_read_audio_stream_info(AVFormatContext *s, AVIOContext *pb,
st->codecpar->block_align = coded_framesize;
break;
case AV_CODEC_ID_COOK:
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
sti->need_parsing = AVSTREAM_PARSE_HEADERS;
case AV_CODEC_ID_ATRAC3:
case AV_CODEC_ID_SIPR:
if (read_all) {
@ -235,7 +236,7 @@ static int rm_read_audio_stream_info(AVFormatContext *s, AVIOContext *pb,
return -1;
}
st->codecpar->block_align = ff_sipr_subpk_size[flavor];
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
sti->need_parsing = AVSTREAM_PARSE_FULL_RAW;
} else {
if(sub_packet_size <= 0){
av_log(s, AV_LOG_ERROR, "sub_packet_size is invalid\n");
@ -388,7 +389,7 @@ int ff_rm_read_mdpr_codecdata(AVFormatContext *s, AVIOContext *pb,
avio_skip(pb, 2); // looks like bits per sample
avio_skip(pb, 4); // always zero?
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->internal->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_TIMESTAMPS;
fps = avio_rb32(pb);
if ((ret = rm_read_extradata(s, pb, st->codecpar, codec_data_size - (avio_tell(pb) - codec_pos))) < 0)

View File

@ -311,6 +311,7 @@ static int rpl_read_packet(AVFormatContext *s, AVPacket *pkt)
RPLContext *rpl = s->priv_data;
AVIOContext *pb = s->pb;
AVStream* stream;
FFStream *sti;
AVIndexEntry* index_entry;
int ret;
@ -320,11 +321,12 @@ static int rpl_read_packet(AVFormatContext *s, AVPacket *pkt)
}
stream = s->streams[rpl->chunk_part];
sti = ffstream(stream);
if (rpl->chunk_number >= stream->internal->nb_index_entries)
if (rpl->chunk_number >= sti->nb_index_entries)
return AVERROR_EOF;
index_entry = &stream->internal->index_entries[rpl->chunk_number];
index_entry = &sti->index_entries[rpl->chunk_number];
if (rpl->frame_in_part == 0) {
if (avio_seek(pb, index_entry->pos, SEEK_SET) < 0)

View File

@ -166,8 +166,8 @@ static int asfrtp_parse_sdp_line(AVFormatContext *s, int stream_index,
if (s->streams[stream_index]->id == rt->asf_ctx->streams[i]->id) {
avcodec_parameters_copy(s->streams[stream_index]->codecpar,
rt->asf_ctx->streams[i]->codecpar);
s->streams[stream_index]->internal->need_parsing =
rt->asf_ctx->streams[i]->internal->need_parsing;
ffstream(s->streams[stream_index])->need_parsing =
ffstream(rt->asf_ctx->streams[i])->need_parsing;
avpriv_set_pts_info(s->streams[stream_index], 32, 1, 1000);
}
}

View File

@ -218,7 +218,7 @@ static void init_rtp_handler(const RTPDynamicProtocolHandler *handler,
par->codec_id = handler->codec_id;
rtsp_st->dynamic_handler = handler;
if (st)
st->internal->need_parsing = handler->need_parsing;
ffstream(st)->need_parsing = handler->need_parsing;
if (handler->priv_data_size) {
rtsp_st->dynamic_protocol_context = av_mallocz(handler->priv_data_size);
if (!rtsp_st->dynamic_protocol_context)

View File

@ -181,7 +181,7 @@ static int s337m_read_packet(AVFormatContext *s, AVPacket *pkt)
}
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = codec;
st->internal->need_parsing = AVSTREAM_PARSE_HEADERS;
ffstream(st)->need_parsing = AVSTREAM_PARSE_HEADERS;
}
return 0;

View File

@ -1409,6 +1409,7 @@ static av_cold int sbg_read_header(AVFormatContext *avf)
char *buf = NULL;
struct sbg_script script = { 0 };
AVStream *st;
FFStream *sti;
struct ws_intervals inter = { 0 };
r = read_whole_file(avf->pb, sbg->max_file_size, &buf);
@ -1442,6 +1443,7 @@ static av_cold int sbg_read_header(AVFormatContext *avf)
st = avformat_new_stream(avf, NULL);
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_FFWAVESYNTH;
st->codecpar->channels = 2;
@ -1449,13 +1451,13 @@ static av_cold int sbg_read_header(AVFormatContext *avf)
st->codecpar->sample_rate = sbg->sample_rate;
st->codecpar->frame_size = sbg->frame_size;
avpriv_set_pts_info(st, 64, 1, st->codecpar->sample_rate);
st->internal->probe_packets = 0;
sti->probe_packets = 0;
st->start_time = av_rescale(script.start_ts,
sbg->sample_rate, AV_TIME_BASE);
st->duration = script.end_ts == AV_NOPTS_VALUE ? AV_NOPTS_VALUE :
av_rescale(script.end_ts - script.start_ts,
sbg->sample_rate, AV_TIME_BASE);
st->internal->cur_dts = st->start_time;
sti->cur_dts = st->start_time;
r = encode_intervals(&script, st->codecpar, &inter);
if (r < 0)
goto fail;
@ -1476,7 +1478,7 @@ static int sbg_read_packet(AVFormatContext *avf, AVPacket *packet)
int64_t ts, end_ts;
int ret;
ts = avf->streams[0]->internal->cur_dts;
ts = ffstream(avf->streams[0])->cur_dts;
end_ts = ts + avf->streams[0]->codecpar->frame_size;
if (avf->streams[0]->duration != AV_NOPTS_VALUE)
end_ts = FFMIN(avf->streams[0]->start_time + avf->streams[0]->duration,
@ -1499,7 +1501,7 @@ static int sbg_read_seek2(AVFormatContext *avf, int stream_index,
return AVERROR(EINVAL);
if (stream_index < 0)
ts = av_rescale_q(ts, AV_TIME_BASE_Q, avf->streams[0]->time_base);
avf->streams[0]->internal->cur_dts = ts;
ffstream(avf->streams[0])->cur_dts = ts;
return 0;
}

View File

@ -51,7 +51,7 @@ static int sdr2_read_header(AVFormatContext *s)
st->codecpar->width = avio_rl32(s->pb);
st->codecpar->height = avio_rl32(s->pb);
st->codecpar->codec_id = AV_CODEC_ID_H264;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
ast->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
ast->codecpar->channels = 1;

View File

@ -183,7 +183,7 @@ static int film_read_header(AVFormatContext *s)
if (film->audio_type == AV_CODEC_ID_ADPCM_ADX) {
st->codecpar->bits_per_coded_sample = 18 * 8 / 32;
st->codecpar->block_align = st->codecpar->channels * 18;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
} else {
st->codecpar->bits_per_coded_sample = film->audio_bits;
st->codecpar->block_align = st->codecpar->channels *
@ -315,7 +315,7 @@ static int film_read_seek(AVFormatContext *s, int stream_index, int64_t timestam
if (ret < 0)
return ret;
pos = avio_seek(s->pb, st->internal->index_entries[ret].pos, SEEK_SET);
pos = avio_seek(s->pb, ffstream(st)->index_entries[ret].pos, SEEK_SET);
if (pos < 0)
return pos;

View File

@ -991,10 +991,10 @@ static int seg_check_bitstream(struct AVFormatContext *s, const AVPacket *pkt)
if (oc->oformat->check_bitstream) {
int ret = oc->oformat->check_bitstream(oc, pkt);
if (ret == 1) {
AVStream *st = s->streams[pkt->stream_index];
AVStream *ost = oc->streams[pkt->stream_index];
st->internal->bsfc = ost->internal->bsfc;
ost->internal->bsfc = NULL;
FFStream *const sti = ffstream( s->streams[pkt->stream_index]);
FFStream *const osti = ffstream(oc->streams[pkt->stream_index]);
sti->bsfc = osti->bsfc;
osti->bsfc = NULL;
}
return ret;
}

View File

@ -582,15 +582,16 @@ static int ism_write_packet(AVFormatContext *s, AVPacket *pkt)
{
SmoothStreamingContext *c = s->priv_data;
AVStream *st = s->streams[pkt->stream_index];
FFStream *const sti = ffstream(st);
OutputStream *os = &c->streams[pkt->stream_index];
int64_t end_dts = (c->nb_fragments + 1) * (int64_t) c->min_frag_duration;
int ret;
if (st->internal->first_dts == AV_NOPTS_VALUE)
st->internal->first_dts = pkt->dts;
if (sti->first_dts == AV_NOPTS_VALUE)
sti->first_dts = pkt->dts;
if ((!c->has_video || st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) &&
av_compare_ts(pkt->dts - st->internal->first_dts, st->time_base,
av_compare_ts(pkt->dts - sti->first_dts, st->time_base,
end_dts, AV_TIME_BASE_Q) >= 0 &&
pkt->flags & AV_PKT_FLAG_KEY && os->packets_written) {

View File

@ -204,7 +204,7 @@ static AVStream *create_new_audio_stream(AVFormatContext *s, int id, int info)
}
ast->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
ast->codecpar->codec_id = ff_codec_get_id(swf_audio_codec_tags, info>>4 & 15);
ast->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(ast)->need_parsing = AVSTREAM_PARSE_FULL;
sample_rate_code = info>>2 & 3;
sample_size_code = info>>1 & 1;
if (!sample_size_code && ast->codecpar->codec_id == AV_CODEC_ID_PCM_S16LE)
@ -293,7 +293,7 @@ static int swf_read_packet(AVFormatContext *s, AVPacket *pkt)
return AVERROR(ENOMEM);
ast->duration = avio_rl32(pb); // number of samples
if (((v>>4) & 15) == 2) { // MP3 sound data record
ast->internal->skip_samples = avio_rl16(pb);
ffstream(ast)->skip_samples = avio_rl16(pb);
len -= 2;
}
len -= 7;

View File

@ -64,7 +64,7 @@ static int tak_read_header(AVFormatContext *s)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_TAK;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
tc->mlast_frame = 0;
if (avio_rl32(pb) != MKTAG('t', 'B', 'a', 'K')) {

View File

@ -277,12 +277,14 @@ static av_cold int tedcaptions_read_header(AVFormatContext *avf)
{
TEDCaptionsDemuxer *tc = avf->priv_data;
AVStream *st = avformat_new_stream(avf, NULL);
FFStream *sti;
int ret, i;
AVPacket *last;
if (!st)
return AVERROR(ENOMEM);
sti = ffstream(st);
ret = parse_file(avf->pb, &tc->subs);
if (ret < 0) {
if (ret == AVERROR_INVALIDDATA)
@ -298,10 +300,10 @@ static av_cold int tedcaptions_read_header(AVFormatContext *avf)
st->codecpar->codec_type = AVMEDIA_TYPE_SUBTITLE;
st->codecpar->codec_id = AV_CODEC_ID_TEXT;
avpriv_set_pts_info(st, 64, 1, 1000);
st->internal->probe_packets = 0;
sti->probe_packets = 0;
st->start_time = 0;
st->duration = last->pts + last->duration;
st->internal->cur_dts = 0;
sti->cur_dts = 0;
return 0;
}

View File

@ -151,21 +151,22 @@ static int tta_read_packet(AVFormatContext *s, AVPacket *pkt)
{
TTAContext *c = s->priv_data;
AVStream *st = s->streams[0];
FFStream *const sti = ffstream(st);
int size, ret;
// FIXME!
if (c->currentframe >= c->totalframes)
return AVERROR_EOF;
if (st->internal->nb_index_entries < c->totalframes) {
if (sti->nb_index_entries < c->totalframes) {
av_log(s, AV_LOG_ERROR, "Index entry disappeared\n");
return AVERROR_INVALIDDATA;
}
size = st->internal->index_entries[c->currentframe].size;
size = sti->index_entries[c->currentframe].size;
ret = av_get_packet(s->pb, pkt, size);
pkt->dts = st->internal->index_entries[c->currentframe++].timestamp;
pkt->dts = sti->index_entries[c->currentframe++].timestamp;
pkt->duration = c->currentframe == c->totalframes ? c->last_frame_size :
c->frame_size;
return ret;
@ -178,7 +179,7 @@ static int tta_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp
int index = av_index_search_timestamp(st, timestamp, flags);
if (index < 0)
return -1;
if (avio_seek(s->pb, st->internal->index_entries[index].pos, SEEK_SET) < 0)
if (avio_seek(s->pb, ffstream(st)->index_entries[index].pos, SEEK_SET) < 0)
return -1;
c->currentframe = index;

View File

@ -308,7 +308,7 @@ static int ty_read_header(AVFormatContext *s)
return AVERROR(ENOMEM);
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
st->codecpar->codec_id = AV_CODEC_ID_MPEG2VIDEO;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 1, 90000);
ast = avformat_new_stream(s, NULL);
@ -318,7 +318,7 @@ static int ty_read_header(AVFormatContext *s)
if (ty->audio_type == TIVO_AUDIO_MPEG) {
ast->codecpar->codec_id = AV_CODEC_ID_MP2;
ast->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(ast)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
} else {
ast->codecpar->codec_id = AV_CODEC_ID_AC3;
}

View File

@ -103,7 +103,7 @@ static int is_relative(int64_t ts) {
*/
static int64_t wrap_timestamp(const AVStream *st, int64_t timestamp)
{
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = cffstream(st);
if (sti->pts_wrap_behavior != AV_PTS_WRAP_IGNORE && st->pts_wrap_bits < 64 &&
sti->pts_wrap_reference != AV_NOPTS_VALUE && timestamp != AV_NOPTS_VALUE) {
if (sti->pts_wrap_behavior == AV_PTS_WRAP_ADD_OFFSET &&
@ -118,20 +118,20 @@ static int64_t wrap_timestamp(const AVStream *st, int64_t timestamp)
int64_t av_stream_get_end_pts(const AVStream *st)
{
if (st->internal->priv_pts) {
return st->internal->priv_pts->val;
if (cffstream(st)->priv_pts) {
return cffstream(st)->priv_pts->val;
} else
return AV_NOPTS_VALUE;
}
struct AVCodecParserContext *av_stream_get_parser(const AVStream *st)
{
return st->internal->parser;
return cffstream(st)->parser;
}
void avpriv_stream_set_need_parsing(AVStream *st, enum AVStreamParseType type)
{
st->internal->need_parsing = type;
ffstream(st)->need_parsing = type;
}
void av_format_inject_global_side_data(AVFormatContext *s)
@ -140,7 +140,7 @@ void av_format_inject_global_side_data(AVFormatContext *s)
si->inject_global_side_data = 1;
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
st->internal->inject_global_side_data = 1;
ffstream(st)->inject_global_side_data = 1;
}
}
@ -312,9 +312,9 @@ static int set_codec_from_probe_data(AVFormatContext *s, AVStream *st,
};
int score;
const AVInputFormat *fmt = av_probe_input_format3(pd, 1, &score);
FFStream *const sti = ffstream(st);
if (fmt) {
AVStreamInternal *const sti = st->internal;
av_log(s, AV_LOG_DEBUG,
"Probe with size=%d, packets=%d detected %s with score=%d\n",
pd->buf_size, s->max_probe_packets - sti->probe_packets,
@ -436,7 +436,7 @@ static int update_stream_avctx(AVFormatContext *s)
int ret;
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (!sti->need_context_update)
continue;
@ -633,7 +633,7 @@ static void force_codec_ids(AVFormatContext *s, AVStream *st)
static int probe_codec(AVFormatContext *s, AVStream *st, const AVPacket *pkt)
{
FFFormatContext *const si = ffformatcontext(s);
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (sti->request_probe > 0) {
AVProbeData *pd = &sti->probe_data;
@ -685,7 +685,7 @@ no_packet:
static int update_wrap_reference(AVFormatContext *s, AVStream *st, int stream_index, AVPacket *pkt)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int64_t ref = pkt->dts;
int pts_wrap_behavior;
int64_t pts_wrap_reference;
@ -708,10 +708,10 @@ static int update_wrap_reference(AVFormatContext *s, AVStream *st, int stream_in
if (!first_program) {
int default_stream_index = av_find_default_stream_index(s);
AVStreamInternal *const default_sti = s->streams[default_stream_index]->internal;
FFStream *const default_sti = ffstream(s->streams[default_stream_index]);
if (default_sti->pts_wrap_reference == AV_NOPTS_VALUE) {
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStreamInternal *const sti = s->streams[i]->internal;
FFStream *const sti = ffstream(s->streams[i]);
if (av_find_program_from_stream(s, NULL, i))
continue;
sti->pts_wrap_reference = pts_wrap_reference;
@ -739,7 +739,7 @@ static int update_wrap_reference(AVFormatContext *s, AVStream *st, int stream_in
while (program) {
if (program->pts_wrap_reference != pts_wrap_reference) {
for (unsigned i = 0; i < program->nb_stream_indexes; i++) {
AVStreamInternal *const sti = s->streams[program->stream_index[i]]->internal;
FFStream *const sti = ffstream(s->streams[program->stream_index[i]]);
sti->pts_wrap_reference = pts_wrap_reference;
sti->pts_wrap_behavior = pts_wrap_behavior;
}
@ -771,7 +771,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
for (;;) {
PacketList *pktl = si->raw_packet_buffer;
AVStreamInternal *sti;
FFStream *sti;
const AVPacket *pkt1;
if (pktl) {
@ -779,7 +779,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
if (si->raw_packet_buffer_remaining_size <= 0)
if ((err = probe_codec(s, st, NULL)) < 0)
return err;
if (st->internal->request_probe <= 0) {
if (ffstream(st)->request_probe <= 0) {
avpriv_packet_list_get(&si->raw_packet_buffer,
&si->raw_packet_buffer_end, pkt);
si->raw_packet_buffer_remaining_size += pkt->size;
@ -800,7 +800,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
return err;
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (sti->probe_packets || sti->request_probe > 0)
if ((err = probe_codec(s, st, NULL)) < 0)
return err;
@ -831,7 +831,7 @@ FF_ENABLE_DEPRECATION_WARNINGS
"Invalid stream index.\n");
st = s->streams[pkt->stream_index];
sti = st->internal;
sti = ffstream(st);
if (update_wrap_reference(s, st, pkt->stream_index, pkt) && sti->pts_wrap_behavior == AV_PTS_WRAP_SUB_OFFSET) {
// correct first time stamps to negative values
@ -893,7 +893,7 @@ static void compute_frame_duration(AVFormatContext *s, int *pnum, int *pden,
AVStream *st, AVCodecParserContext *pc,
AVPacket *pkt)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVRational codec_framerate = sti->avctx->framerate;
int frame_size, sample_rate;
@ -958,7 +958,7 @@ int ff_is_intra_only(enum AVCodecID id)
static int has_decode_delay_been_guessed(AVStream *st)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (st->codecpar->codec_id != AV_CODEC_ID_H264) return 1;
if (!sti->info) // if we have left find_stream_info then nb_decoded_frames won't increase anymore for stream copy
return 1;
@ -986,7 +986,7 @@ static PacketList *get_next_pkt(AVFormatContext *s, AVStream *st, PacketList *pk
}
static int64_t select_from_pts_buffer(AVStream *st, int64_t *pts_buffer, int64_t dts) {
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int onein_oneout = st->codecpar->codec_id != AV_CODEC_ID_H264 &&
st->codecpar->codec_id != AV_CODEC_ID_HEVC;
@ -1035,7 +1035,7 @@ static void update_dts_from_pts(AVFormatContext *s, int stream_index,
PacketList *pkt_buffer)
{
AVStream *st = s->streams[stream_index];
int delay = st->internal->avctx->has_b_frames;
int delay = ffstream(st)->avctx->has_b_frames;
int64_t pts_buffer[MAX_REORDER_DELAY+1];
@ -1061,7 +1061,7 @@ static void update_initial_timestamps(AVFormatContext *s, int stream_index,
{
FFFormatContext *const si = ffformatcontext(s);
AVStream *st = s->streams[stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
PacketList *pktl = si->packet_buffer ? si->packet_buffer : si->parse_queue;
PacketList *pktl_it;
@ -1115,7 +1115,7 @@ static void update_initial_durations(AVFormatContext *s, AVStream *st,
int stream_index, int64_t duration)
{
FFFormatContext *const si = ffformatcontext(s);
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
PacketList *pktl = si->packet_buffer ? si->packet_buffer : si->parse_queue;
int64_t cur_dts = RELATIVE_TS_BASE;
@ -1175,7 +1175,7 @@ static void compute_pkt_fields(AVFormatContext *s, AVStream *st,
int64_t next_dts, int64_t next_pts)
{
FFFormatContext *const si = ffformatcontext(s);
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int num, den, presentation_delayed, delay;
int64_t offset;
AVRational duration;
@ -1369,7 +1369,7 @@ static int parse_packet(AVFormatContext *s, AVPacket *pkt,
FFFormatContext *const si = ffformatcontext(s);
AVPacket *out_pkt = si->parse_pkt;
AVStream *st = s->streams[stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
uint8_t *data = pkt->data;
int size = pkt->size;
int ret = 0, got_output = flush;
@ -1487,7 +1487,7 @@ static int read_frame_internal(AVFormatContext *s, AVPacket *pkt)
while (!got_packet && !si->parse_queue) {
AVStream *st;
AVStreamInternal *sti;
FFStream *sti;
/* read next packet */
ret = ff_read_packet(s, pkt);
@ -1497,7 +1497,8 @@ static int read_frame_internal(AVFormatContext *s, AVPacket *pkt)
/* flush the parsers */
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
if (st->internal->parser && st->internal->need_parsing)
FFStream *const sti = ffstream(st);
if (sti->parser && sti->need_parsing)
parse_packet(s, pkt, st->index, 1);
}
/* all remaining packets are now in parse_queue =>
@ -1506,7 +1507,7 @@ static int read_frame_internal(AVFormatContext *s, AVPacket *pkt)
}
ret = 0;
st = s->streams[pkt->stream_index];
sti = st->internal;
sti = ffstream(st);
st->event_flags |= AVSTREAM_EVENT_FLAG_NEW_PACKETS;
@ -1602,7 +1603,7 @@ static int read_frame_internal(AVFormatContext *s, AVPacket *pkt)
if (ret >= 0) {
AVStream *st = s->streams[pkt->stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int discard_padding = 0;
if (sti->first_discard_sample && pkt->pts != AV_NOPTS_VALUE) {
int64_t pts = pkt->pts - (is_relative(pkt->pts) ? RELATIVE_TS_BASE : 0);
@ -1792,6 +1793,7 @@ int av_find_default_stream_index(AVFormatContext *s)
return -1;
for (unsigned i = 0; i < s->nb_streams; i++) {
const AVStream *const st = s->streams[i];
const FFStream *const sti = cffstream(st);
int score = 0;
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
if (st->disposition & AV_DISPOSITION_ATTACHED_PIC)
@ -1804,7 +1806,7 @@ int av_find_default_stream_index(AVFormatContext *s)
if (st->codecpar->sample_rate)
score += 50;
}
if (st->internal->codec_info_nb_frames)
if (sti->codec_info_nb_frames)
score += 12;
if (st->discard != AVDISCARD_ALL)
@ -1828,7 +1830,7 @@ void ff_read_frame_flush(AVFormatContext *s)
/* Reset read state for each stream. */
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *const st = s->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (sti->parser) {
av_parser_close(sti->parser);
@ -1858,8 +1860,9 @@ void avpriv_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timesta
{
for (unsigned i = 0; i < s->nb_streams; i++) {
AVStream *st = s->streams[i];
FFStream *const sti = ffstream(st);
st->internal->cur_dts =
sti->cur_dts =
av_rescale(timestamp,
st->time_base.den * (int64_t) ref_st->time_base.num,
st->time_base.num * (int64_t) ref_st->time_base.den);
@ -1869,7 +1872,7 @@ void avpriv_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timesta
void ff_reduce_index(AVFormatContext *s, int stream_index)
{
AVStream *st = s->streams[stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
unsigned int max_entries = s->max_index_size / sizeof(AVIndexEntry);
if ((unsigned) sti->nb_index_entries >= max_entries) {
@ -1942,7 +1945,7 @@ int ff_add_index_entry(AVIndexEntry **index_entries,
int av_add_index_entry(AVStream *st, int64_t pos, int64_t timestamp,
int size, int distance, int flags)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
timestamp = wrap_timestamp(st, timestamp);
return ff_add_index_entry(&sti->index_entries, &sti->nb_index_entries,
&sti->index_entries_allocated_size, pos,
@ -2014,10 +2017,10 @@ void ff_configure_buffers_for_index(AVFormatContext *s, int64_t time_tolerance)
for (unsigned ist1 = 0; ist1 < s->nb_streams; ist1++) {
AVStream *st1 = s->streams[ist1];
AVStreamInternal *const sti1 = st1->internal;
FFStream *const sti1 = ffstream(st1);
for (unsigned ist2 = 0; ist2 < s->nb_streams; ist2++) {
AVStream *st2 = s->streams[ist2];
AVStreamInternal *const sti2 = st2->internal;
FFStream *const sti2 = ffstream(st2);
if (ist1 == ist2)
continue;
@ -2061,19 +2064,19 @@ void ff_configure_buffers_for_index(AVFormatContext *s, int64_t time_tolerance)
int av_index_search_timestamp(AVStream *st, int64_t wanted_timestamp, int flags)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
return ff_index_search_timestamp(sti->index_entries, sti->nb_index_entries,
wanted_timestamp, flags);
}
int avformat_index_get_entries_count(const AVStream *st)
{
return st->internal->nb_index_entries;
return cffstream(st)->nb_index_entries;
}
const AVIndexEntry *avformat_index_get_entry(AVStream *st, int idx)
{
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = ffstream(st);
if (idx < 0 || idx >= sti->nb_index_entries)
return NULL;
@ -2084,7 +2087,7 @@ const AVIndexEntry *avformat_index_get_entry_from_timestamp(AVStream *st,
int64_t wanted_timestamp,
int flags)
{
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = ffstream(st);
int idx = ff_index_search_timestamp(sti->index_entries,
sti->nb_index_entries,
wanted_timestamp, flags);
@ -2113,7 +2116,7 @@ int ff_seek_frame_binary(AVFormatContext *s, int stream_index,
int index;
int64_t ret;
AVStream *st;
AVStreamInternal *sti;
FFStream *sti;
if (stream_index < 0)
return -1;
@ -2125,7 +2128,7 @@ int ff_seek_frame_binary(AVFormatContext *s, int stream_index,
pos_limit = -1; // GCC falsely says it may be uninitialized.
st = s->streams[stream_index];
sti = st->internal;
sti = ffstream(st);
if (sti->index_entries) {
AVIndexEntry *e;
@ -2345,7 +2348,7 @@ static int seek_frame_generic(AVFormatContext *s, int stream_index,
{
FFFormatContext *const si = ffformatcontext(s);
AVStream *const st = s->streams[stream_index];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int index;
int64_t ret;
AVIndexEntry *ie;
@ -2687,7 +2690,7 @@ static void estimate_timings_from_bit_rate(AVFormatContext *ic)
int64_t bit_rate = 0;
for (unsigned i = 0; i < ic->nb_streams; i++) {
const AVStream *const st = ic->streams[i];
const AVStreamInternal *const sti = st->internal;
const FFStream *const sti = cffstream(st);
if (st->codecpar->bit_rate <= 0 && sti->avctx->bit_rate > 0)
st->codecpar->bit_rate = sti->avctx->bit_rate;
if (st->codecpar->bit_rate > 0) {
@ -2749,7 +2752,7 @@ static void estimate_timings_from_pts(AVFormatContext *ic, int64_t old_offset)
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (st->start_time == AV_NOPTS_VALUE &&
sti->first_dts == AV_NOPTS_VALUE &&
@ -2782,7 +2785,7 @@ static void estimate_timings_from_pts(AVFormatContext *ic, int64_t old_offset)
read_size = 0;
for (;;) {
AVStream *st;
AVStreamInternal *sti;
FFStream *sti;
if (read_size >= DURATION_MAX_READ_SIZE << (FFMAX(retry - 1, 0)))
break;
@ -2793,7 +2796,7 @@ static void estimate_timings_from_pts(AVFormatContext *ic, int64_t old_offset)
break;
read_size += pkt->size;
st = ic->streams[pkt->stream_index];
sti = st->internal;
sti = ffstream(st);
if (pkt->pts != AV_NOPTS_VALUE &&
(st->start_time != AV_NOPTS_VALUE ||
sti->first_dts != AV_NOPTS_VALUE)) {
@ -2844,12 +2847,13 @@ static void estimate_timings_from_pts(AVFormatContext *ic, int64_t old_offset)
/* warn about audio/video streams which duration could not be estimated */
for (unsigned i = 0; i < ic->nb_streams; i++) {
const AVStream *const st = ic->streams[i];
const FFStream *const sti = cffstream(st);
if (st->duration == AV_NOPTS_VALUE) {
switch (st->codecpar->codec_type) {
case AVMEDIA_TYPE_VIDEO:
case AVMEDIA_TYPE_AUDIO:
if (st->start_time != AV_NOPTS_VALUE || st->internal->first_dts != AV_NOPTS_VALUE) {
if (st->start_time != AV_NOPTS_VALUE || sti->first_dts != AV_NOPTS_VALUE) {
av_log(ic, AV_LOG_WARNING, "stream %d : no PTS found at end of file, duration not set\n", i);
} else
av_log(ic, AV_LOG_WARNING, "stream %d : no TS found at start of file, duration not set\n", i);
@ -2862,7 +2866,7 @@ skip_duration_calc:
avio_seek(ic->pb, old_offset, SEEK_SET);
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
sti->cur_dts = sti->first_dts;
sti->last_IP_pts = AV_NOPTS_VALUE;
@ -2935,7 +2939,7 @@ static void estimate_timings(AVFormatContext *ic, int64_t old_offset)
static int has_codec_parameters(AVStream *st, const char **errmsg_ptr)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVCodecContext *avctx = sti->avctx;
#define FAIL(errmsg) do { \
@ -2985,7 +2989,7 @@ static int has_codec_parameters(AVStream *st, const char **errmsg_ptr)
static int try_decode_frame(AVFormatContext *s, AVStream *st,
const AVPacket *avpkt, AVDictionary **options)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVCodecContext *avctx = sti->avctx;
const AVCodec *codec;
int got_picture = 1, ret = 0;
@ -3313,7 +3317,7 @@ int ff_get_extradata(AVFormatContext *s, AVCodecParameters *par, AVIOContext *pb
int ff_rfps_add_frame(AVFormatContext *ic, AVStream *st, int64_t ts)
{
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int64_t last = sti->info->last_dts;
if ( ts != AV_NOPTS_VALUE && last != AV_NOPTS_VALUE && ts > last
@ -3375,7 +3379,7 @@ void ff_rfps_calculate(AVFormatContext *ic)
{
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (st->codecpar->codec_type != AVMEDIA_TYPE_VIDEO)
continue;
@ -3455,7 +3459,7 @@ static int extract_extradata_check(AVStream *st)
static int extract_extradata_init(AVStream *st)
{
AVStreamInternal *sti = st->internal;
FFStream *sti = ffstream(st);
const AVBitStreamFilter *f;
int ret;
@ -3494,7 +3498,7 @@ fail:
static int extract_extradata(FFFormatContext *si, AVStream *st, const AVPacket *pkt)
{
AVStreamInternal *sti = st->internal;
FFStream *sti = ffstream(st);
AVPacket *pkt_ref = si->parse_pkt;
int ret;
@ -3597,7 +3601,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
const AVCodec *codec;
AVDictionary *thread_opt = NULL;
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVCodecContext *const avctx = sti->avctx;
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO ||
@ -3661,7 +3665,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
for (;;) {
const AVPacket *pkt;
AVStream *st;
AVStreamInternal *sti;
FFStream *sti;
AVCodecContext *avctx;
int analyzed_all_streams;
unsigned i;
@ -3674,7 +3678,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
/* check if one codec still needs to be handled */
for (i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
int fps_analyze_framecount = 20;
int count;
@ -3736,8 +3740,9 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
"Probe buffer size limit of %"PRId64" bytes reached\n", probesize);
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
FFStream *const sti = ffstream(st);
if (!st->r_frame_rate.num &&
st->internal->info->duration_count <= 1 &&
sti->info->duration_count <= 1 &&
st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO &&
strcmp(ic->iformat->name, "image2"))
av_log(ic, AV_LOG_WARNING,
@ -3772,7 +3777,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
}
st = ic->streams[pkt->stream_index];
sti = st->internal;
sti = ffstream(st);
if (!(st->disposition & AV_DISPOSITION_ATTACHED_PIC))
read_size += pkt->size;
@ -3898,7 +3903,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
if (eof_reached) {
for (unsigned stream_index = 0; stream_index < ic->nb_streams; stream_index++) {
AVStream *const st = ic->streams[stream_index];
AVCodecContext *const avctx = st->internal->avctx;
AVCodecContext *const avctx = ffstream(st)->avctx;
if (!has_codec_parameters(st, NULL)) {
const AVCodec *codec = find_probe_decoder(ic, st, st->codecpar->codec_id);
if (codec && !avctx->codec) {
@ -3927,9 +3932,10 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
FFStream *const sti = ffstream(st);
/* flush the decoders */
if (st->internal->info->found_decoder == 1) {
if (sti->info->found_decoder == 1) {
do {
err = try_decode_frame(ic, st, empty_pkt,
(options && i < orig_nb_streams)
@ -3948,7 +3954,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
AVCodecContext *const avctx = sti->avctx;
if (avctx->codec_type == AVMEDIA_TYPE_VIDEO) {
@ -4050,7 +4056,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
ret = -1;
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
const char *errmsg;
/* if no packet was ever seen, update context now for has_codec_parameters */
@ -4081,7 +4087,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
/* update the stream parameters from the internal codec contexts */
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (sti->avctx_inited) {
ret = avcodec_parameters_from_context(st->codecpar, sti->avctx);
@ -4098,7 +4104,7 @@ int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options)
find_stream_info_err:
for (unsigned i = 0; i < ic->nb_streams; i++) {
AVStream *const st = ic->streams[i];
AVStreamInternal *const sti = st->internal;
FFStream *const sti = ffstream(st);
if (sti->info) {
av_freep(&sti->info->duration_error);
av_freep(&sti->info);
@ -4173,7 +4179,7 @@ int av_find_best_stream(AVFormatContext *ic, enum AVMediaType type,
}
disposition = !(st->disposition & (AV_DISPOSITION_HEARING_IMPAIRED | AV_DISPOSITION_VISUAL_IMPAIRED))
+ !! (st->disposition & AV_DISPOSITION_DEFAULT);
count = st->internal->codec_info_nb_frames;
count = ffstream(st)->codec_info_nb_frames;
bitrate = par->bit_rate;
multiframe = FFMIN(5, count);
if ((best_disposition > disposition) ||
@ -4271,7 +4277,7 @@ int ff_stream_encode_params_copy(AVStream *dst, const AVStream *src)
static void free_stream(AVStream **pst)
{
AVStream *st = *pst;
AVStreamInternal *sti;
FFStream *const sti = ffstream(st);
if (!st)
return;
@ -4283,8 +4289,6 @@ static void free_stream(AVStream **pst)
if (st->attached_pic.data)
av_packet_unref(&st->attached_pic);
sti = st->internal;
if (sti) {
av_parser_close(sti->parser);
avcodec_free_context(&sti->avctx);
av_bsf_free(&sti->bsfc);
@ -4298,8 +4302,6 @@ static void free_stream(AVStream **pst)
av_freep(&sti->info->duration_error);
av_freep(&sti->info);
}
}
av_freep(&st->internal);
av_dict_free(&st->metadata);
avcodec_parameters_free(&st->codecpar);
@ -4390,8 +4392,8 @@ void avformat_close_input(AVFormatContext **ps)
AVStream *avformat_new_stream(AVFormatContext *s, const AVCodec *c)
{
FFFormatContext *const si = ffformatcontext(s);
FFStream *sti;
AVStream *st;
AVStreamInternal *sti;
AVStream **streams;
if (s->nb_streams >= s->max_streams) {
@ -4405,13 +4407,11 @@ AVStream *avformat_new_stream(AVFormatContext *s, const AVCodec *c)
return NULL;
s->streams = streams;
st = av_mallocz(sizeof(AVStream));
if (!st)
return NULL;
sti = st->internal = av_mallocz(sizeof(*st->internal));
sti = av_mallocz(sizeof(*sti));
if (!sti)
goto fail;
return NULL;
st = &sti->pub;
st->codecpar = avcodec_parameters_alloc();
if (!st->codecpar)
@ -4814,6 +4814,7 @@ int ff_hex_to_data(uint8_t *data, const char *p)
void avpriv_set_pts_info(AVStream *s, int pts_wrap_bits,
unsigned int pts_num, unsigned int pts_den)
{
FFStream *const sti = ffstream(s);
AVRational new_tb;
if (av_reduce(&new_tb.num, &new_tb.den, pts_num, pts_den, INT_MAX)) {
if (new_tb.num != pts_num)
@ -4832,7 +4833,7 @@ void avpriv_set_pts_info(AVStream *s, int pts_wrap_bits,
return;
}
s->time_base = new_tb;
s->internal->avctx->pkt_timebase = new_tb;
sti->avctx->pkt_timebase = new_tb;
s->pts_wrap_bits = pts_wrap_bits;
}
@ -5005,7 +5006,7 @@ AVRational av_guess_sample_aspect_ratio(AVFormatContext *format, AVStream *strea
AVRational av_guess_frame_rate(AVFormatContext *format, AVStream *st, AVFrame *frame)
{
AVRational fr = st->r_frame_rate;
AVCodecContext *const avctx = st->internal->avctx;
AVCodecContext *const avctx = ffstream(st)->avctx;
AVRational codec_fr = avctx->framerate;
AVRational avg_fr = st->avg_frame_rate;
@ -5407,9 +5408,10 @@ int ff_stream_add_bitstream_filter(AVStream *st, const char *name, const char *a
{
int ret;
const AVBitStreamFilter *bsf;
FFStream *const sti = ffstream(st);
AVBSFContext *bsfc;
av_assert0(!st->internal->bsfc);
av_assert0(!sti->bsfc);
if (!(bsf = av_bsf_get_by_name(name))) {
av_log(NULL, AV_LOG_ERROR, "Unknown bitstream filter '%s'\n", name);
@ -5443,7 +5445,7 @@ int ff_stream_add_bitstream_filter(AVStream *st, const char *name, const char *a
return ret;
}
st->internal->bsfc = bsfc;
sti->bsfc = bsfc;
av_log(NULL, AV_LOG_VERBOSE,
"Automatically inserted bitstream filter '%s'; args='%s'\n",
@ -5550,8 +5552,8 @@ int avformat_transfer_internal_stream_timing_info(const AVOutputFormat *ofmt,
AVStream *ost, const AVStream *ist,
enum AVTimebaseSource copy_tb)
{
const AVCodecContext *const dec_ctx = ist->internal->avctx;
AVCodecContext *const enc_ctx = ost->internal->avctx;
const AVCodecContext *const dec_ctx = cffstream(ist)->avctx;
AVCodecContext *const enc_ctx = ffstream(ost)->avctx;
enc_ctx->time_base = ist->time_base;
/*
@ -5607,7 +5609,7 @@ int avformat_transfer_internal_stream_timing_info(const AVOutputFormat *ofmt,
AVRational av_stream_get_codec_timebase(const AVStream *st)
{
// See avformat_transfer_internal_stream_timing_info() TODO.
return st->internal->avctx->time_base;
return cffstream(st)->avctx->time_base;
}
void ff_format_set_url(AVFormatContext *s, char *url)

View File

@ -73,6 +73,7 @@ static int voc_read_seek(AVFormatContext *s, int stream_index,
{
VocDecContext *voc = s->priv_data;
AVStream *st;
FFStream *sti;
int index;
if (s->nb_streams < 1) {
@ -81,16 +82,17 @@ static int voc_read_seek(AVFormatContext *s, int stream_index,
}
st = s->streams[stream_index];
sti = ffstream(st);
index = av_index_search_timestamp(st, timestamp, flags);
if (index >= 0 && index < st->internal->nb_index_entries - 1) {
AVIndexEntry *e = &st->internal->index_entries[index];
if (index >= 0 && index < sti->nb_index_entries - 1) {
const AVIndexEntry *const e = &sti->index_entries[index];
avio_seek(s->pb, e->pos, SEEK_SET);
voc->pts = e->timestamp;
voc->remaining_size = e->size;
return 0;
} else if (st->internal->nb_index_entries && st->internal->index_entries[0].timestamp <= timestamp) {
AVIndexEntry *e = &st->internal->index_entries[st->internal->nb_index_entries - 1];
} else if (sti->nb_index_entries && sti->index_entries[0].timestamp <= timestamp) {
const AVIndexEntry *const e = &sti->index_entries[sti->nb_index_entries - 1];
// prepare context for seek_frame_generic()
voc->pts = e->timestamp;
voc->remaining_size = e->size;

View File

@ -277,7 +277,7 @@ static int vqf_read_seek(AVFormatContext *s,
AV_ROUND_DOWN : AV_ROUND_UP);
pos *= c->frame_bit_len;
st->internal->cur_dts = av_rescale(pos, st->time_base.den,
ffstream(st)->cur_dts = av_rescale(pos, st->time_base.den,
st->codecpar->bit_rate * (int64_t)st->time_base.num);
if ((ret = avio_seek(s->pb, ((pos-7) >> 3) + ffformatcontext(s)->data_offset, SEEK_SET)) < 0)

View File

@ -166,8 +166,9 @@ static int wav_probe(const AVProbeData *p)
static void handle_stream_probing(AVStream *st)
{
if (st->codecpar->codec_id == AV_CODEC_ID_PCM_S16LE) {
st->internal->request_probe = AVPROBE_SCORE_EXTENSION;
st->internal->probe_packets = FFMIN(st->internal->probe_packets, 32);
FFStream *const sti = ffstream(st);
sti->request_probe = AVPROBE_SCORE_EXTENSION;
sti->probe_packets = FFMIN(sti->probe_packets, 32);
}
}
@ -183,7 +184,7 @@ static int wav_parse_fmt_tag(AVFormatContext *s, int64_t size, AVStream *st)
return ret;
handle_stream_probing(st);
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avpriv_set_pts_info(st, 64, 1, st->codecpar->sample_rate);
@ -200,7 +201,7 @@ static int wav_parse_xma2_tag(AVFormatContext *s, int64_t size, AVStream *st)
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
st->codecpar->codec_id = AV_CODEC_ID_XMA2;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
version = avio_r8(pb);
if (version != 3 && version != 4)
@ -702,8 +703,8 @@ static int wav_read_packet(AVFormatContext *s, AVPacket *pkt)
int64_t audio_dts, video_dts;
AVStream *vst = wav->vst;
smv_retry:
audio_dts = (int32_t)st->internal->cur_dts;
video_dts = (int32_t)vst->internal->cur_dts;
audio_dts = (int32_t)ffstream( st)->cur_dts;
video_dts = (int32_t)ffstream(vst)->cur_dts;
if (audio_dts != AV_NOPTS_VALUE && video_dts != AV_NOPTS_VALUE) {
/*We always return a video frame first to get the pixel format first*/
@ -950,7 +951,7 @@ static int w64_read_header(AVFormatContext *s)
ff_metadata_conv_ctx(s, NULL, ff_riff_info_conv);
handle_stream_probing(st);
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
avio_seek(pb, data_ofs, SEEK_SET);

View File

@ -618,7 +618,7 @@ static AVStream * new_stream(AVFormatContext *s, AVStream *st, int sid, int code
st->priv_data = wst;
}
st->codecpar->codec_type = codec_type;
st->internal->need_parsing = AVSTREAM_PARSE_FULL;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL;
avpriv_set_pts_info(st, 64, 1, 10000000);
return st;
}

View File

@ -87,7 +87,7 @@ static int xvag_read_header(AVFormatContext *s)
if (avio_rb16(s->pb) == 0xFFFB) {
st->codecpar->codec_id = AV_CODEC_ID_MP3;
st->codecpar->block_align = 0x1000;
st->internal->need_parsing = AVSTREAM_PARSE_FULL_RAW;
ffstream(st)->need_parsing = AVSTREAM_PARSE_FULL_RAW;
}
avio_skip(s->pb, -2);

View File

@ -78,7 +78,7 @@ static int xwma_read_header(AVFormatContext *s)
ret = ff_get_wav_header(s, pb, st->codecpar, size, 0);
if (ret < 0)
return ret;
st->internal->need_parsing = AVSTREAM_PARSE_NONE;
ffstream(st)->need_parsing = AVSTREAM_PARSE_NONE;
/* XWMA encoder only allows a few channel/sample rate/bitrate combinations,
* but some create identical files with fake bitrate (1ch 22050hz at