1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-23 12:43:46 +02:00

Implement RMStream stream-specific private data object in the RM demuxer.

This allows multiple video or audio streams per .rm file. See mailinglist
thread "[PATCH] rmdec.c: implement RMVideo/AudioStream".

Originally committed as revision 16365 to svn://svn.ffmpeg.org/ffmpeg/trunk
This commit is contained in:
Ronald S. Bultje 2008-12-28 00:18:38 +00:00
parent 23ae2b68b0
commit ba61472911
2 changed files with 106 additions and 74 deletions

View File

@ -24,6 +24,11 @@
#include "avformat.h" #include "avformat.h"
typedef struct RMStream RMStream;
RMStream *ff_rm_alloc_rmstream (void);
void ff_rm_free_rmstream (RMStream *rms);
/*< input format for Realmedia-style RTSP streams */ /*< input format for Realmedia-style RTSP streams */
extern AVInputFormat rdt_demuxer; extern AVInputFormat rdt_demuxer;

View File

@ -21,12 +21,9 @@
#include "libavutil/avstring.h" #include "libavutil/avstring.h"
#include "avformat.h" #include "avformat.h"
#include "rm.h"
typedef struct { struct RMStream {
int nb_packets;
int old_format;
int current_stream;
int remaining_len;
uint8_t *videobuf; ///< place to store merged video frame uint8_t *videobuf; ///< place to store merged video frame
int videobufsize; ///< current assembled frame size int videobufsize; ///< current assembled frame size
int videobufpos; ///< position for the next slice in the video buffer int videobufpos; ///< position for the next slice in the video buffer
@ -38,10 +35,17 @@ typedef struct {
int64_t audiotimestamp; ///< Audio packet timestamp int64_t audiotimestamp; ///< Audio packet timestamp
int sub_packet_cnt; // Subpacket counter, used while reading int sub_packet_cnt; // Subpacket counter, used while reading
int sub_packet_size, sub_packet_h, coded_framesize; ///< Descrambling parameters from container int sub_packet_size, sub_packet_h, coded_framesize; ///< Descrambling parameters from container
int audio_framesize; /// Audio frame size from container
int sub_packet_lengths[16]; /// Length of each subpacket
};
typedef struct {
int nb_packets;
int old_format;
int current_stream;
int remaining_len;
int audio_stream_num; ///< Stream number for audio packets int audio_stream_num; ///< Stream number for audio packets
int audio_pkt_cnt; ///< Output packet counter int audio_pkt_cnt; ///< Output packet counter
int audio_framesize; /// Audio frame size from container
int sub_packet_lengths[16]; /// Length of each aac subpacket
} RMDemuxContext; } RMDemuxContext;
static inline void get_strl(ByteIOContext *pb, char *buf, int buf_size, int len) static inline void get_strl(ByteIOContext *pb, char *buf, int buf_size, int len)
@ -68,10 +72,24 @@ static void get_str8(ByteIOContext *pb, char *buf, int buf_size)
get_strl(pb, buf, buf_size, get_byte(pb)); get_strl(pb, buf, buf_size, get_byte(pb));
} }
RMStream *ff_rm_alloc_rmstream (void)
{
RMStream *rms = av_mallocz(sizeof(RMStream));
rms->curpic_num = -1;
return rms;
}
void ff_rm_free_rmstream (RMStream *rms)
{
av_free(rms->videobuf);
av_free(rms->audiobuf);
av_free(rms);
}
static int rm_read_audio_stream_info(AVFormatContext *s, ByteIOContext *pb, static int rm_read_audio_stream_info(AVFormatContext *s, ByteIOContext *pb,
AVStream *st, int read_all) AVStream *st, int read_all)
{ {
RMDemuxContext *rm = s->priv_data; RMStream *ast = st->priv_data;
char buf[256]; char buf[256];
uint32_t version; uint32_t version;
@ -104,13 +122,13 @@ static int rm_read_audio_stream_info(AVFormatContext *s, ByteIOContext *pb,
get_be16(pb); /* version2 */ get_be16(pb); /* version2 */
get_be32(pb); /* header size */ get_be32(pb); /* header size */
flavor= get_be16(pb); /* add codec info / flavor */ flavor= get_be16(pb); /* add codec info / flavor */
rm->coded_framesize = coded_framesize = get_be32(pb); /* coded frame size */ ast->coded_framesize = coded_framesize = get_be32(pb); /* coded frame size */
get_be32(pb); /* ??? */ get_be32(pb); /* ??? */
get_be32(pb); /* ??? */ get_be32(pb); /* ??? */
get_be32(pb); /* ??? */ get_be32(pb); /* ??? */
rm->sub_packet_h = sub_packet_h = get_be16(pb); /* 1 */ ast->sub_packet_h = sub_packet_h = get_be16(pb); /* 1 */
st->codec->block_align= get_be16(pb); /* frame size */ st->codec->block_align= get_be16(pb); /* frame size */
rm->sub_packet_size = sub_packet_size = get_be16(pb); /* sub packet size */ ast->sub_packet_size = sub_packet_size = get_be16(pb); /* sub packet size */
get_be16(pb); /* ??? */ get_be16(pb); /* ??? */
if (((version >> 16) & 0xff) == 5) { if (((version >> 16) & 0xff) == 5) {
get_be16(pb); get_be16(pb); get_be16(pb); get_be16(pb); get_be16(pb); get_be16(pb);
@ -133,15 +151,15 @@ static int rm_read_audio_stream_info(AVFormatContext *s, ByteIOContext *pb,
} else if (!strcmp(buf, "28_8")) { } else if (!strcmp(buf, "28_8")) {
st->codec->codec_id = CODEC_ID_RA_288; st->codec->codec_id = CODEC_ID_RA_288;
st->codec->extradata_size= 0; st->codec->extradata_size= 0;
rm->audio_framesize = st->codec->block_align; ast->audio_framesize = st->codec->block_align;
st->codec->block_align = coded_framesize; st->codec->block_align = coded_framesize;
if(rm->audio_framesize >= UINT_MAX / sub_packet_h){ if(ast->audio_framesize >= UINT_MAX / sub_packet_h){
av_log(s, AV_LOG_ERROR, "rm->audio_framesize * sub_packet_h too large\n"); av_log(s, AV_LOG_ERROR, "ast->audio_framesize * sub_packet_h too large\n");
return -1; return -1;
} }
rm->audiobuf = av_malloc(rm->audio_framesize * sub_packet_h); ast->audiobuf = av_malloc(ast->audio_framesize * sub_packet_h);
} else if ((!strcmp(buf, "cook")) || (!strcmp(buf, "atrc")) || (!strcmp(buf, "sipr"))) { } else if ((!strcmp(buf, "cook")) || (!strcmp(buf, "atrc")) || (!strcmp(buf, "sipr"))) {
int codecdata_length; int codecdata_length;
get_be16(pb); get_byte(pb); get_be16(pb); get_byte(pb);
@ -164,15 +182,15 @@ static int rm_read_audio_stream_info(AVFormatContext *s, ByteIOContext *pb,
st->codec->extradata_size= codecdata_length; st->codec->extradata_size= codecdata_length;
st->codec->extradata= av_mallocz(st->codec->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE); st->codec->extradata= av_mallocz(st->codec->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
get_buffer(pb, st->codec->extradata, st->codec->extradata_size); get_buffer(pb, st->codec->extradata, st->codec->extradata_size);
rm->audio_framesize = st->codec->block_align; ast->audio_framesize = st->codec->block_align;
st->codec->block_align = rm->sub_packet_size; st->codec->block_align = ast->sub_packet_size;
if(rm->audio_framesize >= UINT_MAX / sub_packet_h){ if(ast->audio_framesize >= UINT_MAX / sub_packet_h){
av_log(s, AV_LOG_ERROR, "rm->audio_framesize * sub_packet_h too large\n"); av_log(s, AV_LOG_ERROR, "rm->audio_framesize * sub_packet_h too large\n");
return -1; return -1;
} }
rm->audiobuf = av_malloc(rm->audio_framesize * sub_packet_h); ast->audiobuf = av_malloc(ast->audio_framesize * sub_packet_h);
} else if (!strcmp(buf, "raac") || !strcmp(buf, "racp")) { } else if (!strcmp(buf, "raac") || !strcmp(buf, "racp")) {
int codecdata_length; int codecdata_length;
get_be16(pb); get_byte(pb); get_be16(pb); get_byte(pb);
@ -215,7 +233,9 @@ ff_rm_read_mdpr_codecdata (AVFormatContext *s, ByteIOContext *pb,
unsigned int v; unsigned int v;
int size; int size;
int64_t codec_pos; int64_t codec_pos;
RMStream *rst;
st->priv_data = rst = ff_rm_alloc_rmstream();
av_set_pts_info(st, 64, 1, 1000); av_set_pts_info(st, 64, 1, 1000);
codec_pos = url_ftell(pb); codec_pos = url_ftell(pb);
v = get_be32(pb); v = get_be32(pb);
@ -386,7 +406,6 @@ static int rm_read_header(AVFormatContext *s, AVFormatParameters *ap)
if (!rm->nb_packets && (flags & 4)) if (!rm->nb_packets && (flags & 4))
rm->nb_packets = 3600 * 25; rm->nb_packets = 3600 * 25;
get_be32(pb); /* next data header */ get_be32(pb); /* next data header */
rm->curpic_num = -1;
return 0; return 0;
} }
@ -466,7 +485,8 @@ skip:
} }
static int rm_assemble_video_frame(AVFormatContext *s, ByteIOContext *pb, static int rm_assemble_video_frame(AVFormatContext *s, ByteIOContext *pb,
RMDemuxContext *rm, AVPacket *pkt, int len) RMDemuxContext *rm, RMStream *vst,
AVPacket *pkt, int len)
{ {
int hdr, seq, pic_num, len2, pos; int hdr, seq, pic_num, len2, pos;
int type; int type;
@ -507,40 +527,40 @@ static int rm_assemble_video_frame(AVFormatContext *s, ByteIOContext *pb,
} }
//now we have to deal with single slice //now we have to deal with single slice
if((seq & 0x7F) == 1 || rm->curpic_num != pic_num){ if((seq & 0x7F) == 1 || vst->curpic_num != pic_num){
rm->slices = ((hdr & 0x3F) << 1) + 1; vst->slices = ((hdr & 0x3F) << 1) + 1;
rm->videobufsize = len2 + 8*rm->slices + 1; vst->videobufsize = len2 + 8*vst->slices + 1;
av_free(rm->videobuf); av_free(vst->videobuf);
if(!(rm->videobuf = av_malloc(rm->videobufsize))) if(!(vst->videobuf = av_malloc(vst->videobufsize)))
return AVERROR(ENOMEM); return AVERROR(ENOMEM);
rm->videobufpos = 8*rm->slices + 1; vst->videobufpos = 8*vst->slices + 1;
rm->cur_slice = 0; vst->cur_slice = 0;
rm->curpic_num = pic_num; vst->curpic_num = pic_num;
rm->pktpos = url_ftell(pb); vst->pktpos = url_ftell(pb);
} }
if(type == 2) if(type == 2)
len = FFMIN(len, pos); len = FFMIN(len, pos);
if(++rm->cur_slice > rm->slices) if(++vst->cur_slice > vst->slices)
return 1; return 1;
AV_WL32(rm->videobuf - 7 + 8*rm->cur_slice, 1); AV_WL32(vst->videobuf - 7 + 8*vst->cur_slice, 1);
AV_WL32(rm->videobuf - 3 + 8*rm->cur_slice, rm->videobufpos - 8*rm->slices - 1); AV_WL32(vst->videobuf - 3 + 8*vst->cur_slice, vst->videobufpos - 8*vst->slices - 1);
if(rm->videobufpos + len > rm->videobufsize) if(vst->videobufpos + len > vst->videobufsize)
return 1; return 1;
if (get_buffer(pb, rm->videobuf + rm->videobufpos, len) != len) if (get_buffer(pb, vst->videobuf + vst->videobufpos, len) != len)
return AVERROR(EIO); return AVERROR(EIO);
rm->videobufpos += len; vst->videobufpos += len;
rm->remaining_len-= len; rm->remaining_len-= len;
if(type == 2 || (rm->videobufpos) == rm->videobufsize){ if(type == 2 || (vst->videobufpos) == vst->videobufsize){
rm->videobuf[0] = rm->cur_slice-1; vst->videobuf[0] = vst->cur_slice-1;
if(av_new_packet(pkt, rm->videobufpos - 8*(rm->slices - rm->cur_slice)) < 0) if(av_new_packet(pkt, vst->videobufpos - 8*(vst->slices - vst->cur_slice)) < 0)
return AVERROR(ENOMEM); return AVERROR(ENOMEM);
memcpy(pkt->data, rm->videobuf, 1 + 8*rm->cur_slice); memcpy(pkt->data, vst->videobuf, 1 + 8*vst->cur_slice);
memcpy(pkt->data + 1 + 8*rm->cur_slice, rm->videobuf + 1 + 8*rm->slices, memcpy(pkt->data + 1 + 8*vst->cur_slice, vst->videobuf + 1 + 8*vst->slices,
rm->videobufpos - 1 - 8*rm->slices); vst->videobufpos - 1 - 8*vst->slices);
pkt->pts = AV_NOPTS_VALUE; pkt->pts = AV_NOPTS_VALUE;
pkt->pos = rm->pktpos; pkt->pos = vst->pktpos;
return 0; return 0;
} }
@ -571,59 +591,61 @@ ff_rm_parse_packet (AVFormatContext *s, ByteIOContext *pb,
if (st->codec->codec_type == CODEC_TYPE_VIDEO) { if (st->codec->codec_type == CODEC_TYPE_VIDEO) {
rm->current_stream= st->id; rm->current_stream= st->id;
if(rm_assemble_video_frame(s, pb, rm, pkt, len) == 1) if(rm_assemble_video_frame(s, pb, rm, st->priv_data, pkt, len) == 1)
return -1; //got partial frame return -1; //got partial frame
} else if (st->codec->codec_type == CODEC_TYPE_AUDIO) { } else if (st->codec->codec_type == CODEC_TYPE_AUDIO) {
RMStream *ast = st->priv_data;
if ((st->codec->codec_id == CODEC_ID_RA_288) || if ((st->codec->codec_id == CODEC_ID_RA_288) ||
(st->codec->codec_id == CODEC_ID_COOK) || (st->codec->codec_id == CODEC_ID_COOK) ||
(st->codec->codec_id == CODEC_ID_ATRAC3) || (st->codec->codec_id == CODEC_ID_ATRAC3) ||
(st->codec->codec_id == CODEC_ID_SIPR)) { (st->codec->codec_id == CODEC_ID_SIPR)) {
int x; int x;
int sps = rm->sub_packet_size; int sps = ast->sub_packet_size;
int cfs = rm->coded_framesize; int cfs = ast->coded_framesize;
int h = rm->sub_packet_h; int h = ast->sub_packet_h;
int y = rm->sub_packet_cnt; int y = ast->sub_packet_cnt;
int w = rm->audio_framesize; int w = ast->audio_framesize;
if (*flags & 2) if (*flags & 2)
y = rm->sub_packet_cnt = 0; y = ast->sub_packet_cnt = 0;
if (!y) if (!y)
rm->audiotimestamp = *timestamp; ast->audiotimestamp = *timestamp;
switch(st->codec->codec_id) { switch(st->codec->codec_id) {
case CODEC_ID_RA_288: case CODEC_ID_RA_288:
for (x = 0; x < h/2; x++) for (x = 0; x < h/2; x++)
get_buffer(pb, rm->audiobuf+x*2*w+y*cfs, cfs); get_buffer(pb, ast->audiobuf+x*2*w+y*cfs, cfs);
break; break;
case CODEC_ID_ATRAC3: case CODEC_ID_ATRAC3:
case CODEC_ID_COOK: case CODEC_ID_COOK:
for (x = 0; x < w/sps; x++) for (x = 0; x < w/sps; x++)
get_buffer(pb, rm->audiobuf+sps*(h*x+((h+1)/2)*(y&1)+(y>>1)), sps); get_buffer(pb, ast->audiobuf+sps*(h*x+((h+1)/2)*(y&1)+(y>>1)), sps);
break; break;
} }
if (++(rm->sub_packet_cnt) < h) if (++(ast->sub_packet_cnt) < h)
return -1; return -1;
else { else {
rm->sub_packet_cnt = 0; ast->sub_packet_cnt = 0;
rm->audio_stream_num = st->index; rm->audio_stream_num = st->index;
rm->audio_pkt_cnt = h * w / st->codec->block_align - 1; rm->audio_pkt_cnt = h * w / st->codec->block_align - 1;
// Release first audio packet // Release first audio packet
av_new_packet(pkt, st->codec->block_align); av_new_packet(pkt, st->codec->block_align);
memcpy(pkt->data, rm->audiobuf, st->codec->block_align); memcpy(pkt->data, ast->audiobuf, st->codec->block_align);
*timestamp = rm->audiotimestamp; *timestamp = ast->audiotimestamp;
*flags = 2; // Mark first packet as keyframe *flags = 2; // Mark first packet as keyframe
} }
} else if (st->codec->codec_id == CODEC_ID_AAC) { } else if (st->codec->codec_id == CODEC_ID_AAC) {
int x; int x;
rm->audio_stream_num = st->index; rm->audio_stream_num = st->index;
rm->sub_packet_cnt = (get_be16(pb) & 0xf0) >> 4; ast->sub_packet_cnt = (get_be16(pb) & 0xf0) >> 4;
if (rm->sub_packet_cnt) { if (ast->sub_packet_cnt) {
for (x = 0; x < rm->sub_packet_cnt; x++) for (x = 0; x < ast->sub_packet_cnt; x++)
rm->sub_packet_lengths[x] = get_be16(pb); ast->sub_packet_lengths[x] = get_be16(pb);
// Release first audio packet // Release first audio packet
rm->audio_pkt_cnt = rm->sub_packet_cnt - 1; rm->audio_pkt_cnt = ast->sub_packet_cnt - 1;
av_get_packet(pb, pkt, rm->sub_packet_lengths[0]); av_get_packet(pb, pkt, ast->sub_packet_lengths[0]);
*flags = 2; // Mark first packet as keyframe *flags = 2; // Mark first packet as keyframe
} }
} else { } else {
@ -666,15 +688,16 @@ ff_rm_retrieve_cache (AVFormatContext *s, ByteIOContext *pb,
AVStream *st, AVPacket *pkt) AVStream *st, AVPacket *pkt)
{ {
RMDemuxContext *rm = s->priv_data; RMDemuxContext *rm = s->priv_data;
RMStream *ast = st->priv_data;
assert (rm->audio_pkt_cnt > 0); assert (rm->audio_pkt_cnt > 0);
if (st->codec->codec_id == CODEC_ID_AAC) if (st->codec->codec_id == CODEC_ID_AAC)
av_get_packet(pb, pkt, rm->sub_packet_lengths[rm->sub_packet_cnt - rm->audio_pkt_cnt]); av_get_packet(pb, pkt, ast->sub_packet_lengths[ast->sub_packet_cnt - rm->audio_pkt_cnt]);
else { else {
av_new_packet(pkt, st->codec->block_align); av_new_packet(pkt, st->codec->block_align);
memcpy(pkt->data, rm->audiobuf + st->codec->block_align * memcpy(pkt->data, ast->audiobuf + st->codec->block_align *
(rm->sub_packet_h * rm->audio_framesize / st->codec->block_align - rm->audio_pkt_cnt), (ast->sub_packet_h * ast->audio_framesize / st->codec->block_align - rm->audio_pkt_cnt),
st->codec->block_align); st->codec->block_align);
} }
rm->audio_pkt_cnt--; rm->audio_pkt_cnt--;
@ -696,19 +719,22 @@ static int rm_read_packet(AVFormatContext *s, AVPacket *pkt)
st = s->streams[rm->audio_stream_num]; st = s->streams[rm->audio_stream_num];
ff_rm_retrieve_cache(s, s->pb, st, pkt); ff_rm_retrieve_cache(s, s->pb, st, pkt);
} else if (rm->old_format) { } else if (rm->old_format) {
RMStream *ast;
st = s->streams[0]; st = s->streams[0];
ast = st->priv_data;
if (st->codec->codec_id == CODEC_ID_RA_288) { if (st->codec->codec_id == CODEC_ID_RA_288) {
int x, y; int x, y;
for (y = 0; y < rm->sub_packet_h; y++) for (y = 0; y < ast->sub_packet_h; y++)
for (x = 0; x < rm->sub_packet_h/2; x++) for (x = 0; x < ast->sub_packet_h/2; x++)
if (get_buffer(pb, rm->audiobuf+x*2*rm->audio_framesize+y*rm->coded_framesize, rm->coded_framesize) <= 0) if (get_buffer(pb, ast->audiobuf+x*2*ast->audio_framesize+y*ast->coded_framesize, ast->coded_framesize) <= 0)
return AVERROR(EIO); return AVERROR(EIO);
rm->audio_stream_num = 0; rm->audio_stream_num = 0;
rm->audio_pkt_cnt = rm->sub_packet_h * rm->audio_framesize / st->codec->block_align - 1; rm->audio_pkt_cnt = ast->sub_packet_h * ast->audio_framesize / st->codec->block_align - 1;
// Release first audio packet // Release first audio packet
av_new_packet(pkt, st->codec->block_align); av_new_packet(pkt, st->codec->block_align);
memcpy(pkt->data, rm->audiobuf, st->codec->block_align); memcpy(pkt->data, ast->audiobuf, st->codec->block_align);
pkt->flags |= PKT_FLAG_KEY; // Mark first packet as keyframe pkt->flags |= PKT_FLAG_KEY; // Mark first packet as keyframe
pkt->stream_index = 0; pkt->stream_index = 0;
} else { } else {
@ -742,10 +768,11 @@ resync:
static int rm_read_close(AVFormatContext *s) static int rm_read_close(AVFormatContext *s)
{ {
RMDemuxContext *rm = s->priv_data; int i;
for (i=0;i<s->nb_streams;i++)
ff_rm_free_rmstream(s->streams[i]->priv_data);
av_free(rm->audiobuf);
av_free(rm->videobuf);
return 0; return 0;
} }