mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2025-01-13 21:28:01 +02:00
fftools/ffmpeg: move decoding code to ffmpeg_dec
This commit is contained in:
parent
79c1dde5c3
commit
dbf1c6f5f1
534
fftools/ffmpeg.c
534
fftools/ffmpeg.c
@ -283,20 +283,6 @@ static void sub2video_heartbeat(InputStream *ist, int64_t pts)
|
||||
}
|
||||
}
|
||||
|
||||
static void sub2video_flush(InputStream *ist)
|
||||
{
|
||||
int i;
|
||||
int ret;
|
||||
|
||||
if (ist->sub2video.end_pts < INT64_MAX)
|
||||
sub2video_update(ist, INT64_MAX, NULL);
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
ret = av_buffersrc_add_frame(ist->filters[i]->filter, NULL);
|
||||
if (ret != AVERROR_EOF && ret < 0)
|
||||
av_log(NULL, AV_LOG_WARNING, "Flush the frame error.\n");
|
||||
}
|
||||
}
|
||||
|
||||
/* end of sub2video hack */
|
||||
|
||||
static void term_exit_sigsafe(void)
|
||||
@ -768,417 +754,6 @@ static void print_report(int is_last_report, int64_t timer_start, int64_t cur_ti
|
||||
first_report = 0;
|
||||
}
|
||||
|
||||
static void check_decode_result(InputStream *ist, int *got_output, int ret)
|
||||
{
|
||||
if (ret < 0)
|
||||
ist->decode_errors++;
|
||||
|
||||
if (ret < 0 && exit_on_error)
|
||||
exit_program(1);
|
||||
|
||||
if (*got_output && ist->dec_ctx->codec_type != AVMEDIA_TYPE_SUBTITLE) {
|
||||
if (ist->decoded_frame->decode_error_flags || (ist->decoded_frame->flags & AV_FRAME_FLAG_CORRUPT)) {
|
||||
av_log(ist, exit_on_error ? AV_LOG_FATAL : AV_LOG_WARNING,
|
||||
"corrupt decoded frame\n");
|
||||
if (exit_on_error)
|
||||
exit_program(1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// This does not quite work like avcodec_decode_audio4/avcodec_decode_video2.
|
||||
// There is the following difference: if you got a frame, you must call
|
||||
// it again with pkt=NULL. pkt==NULL is treated differently from pkt->size==0
|
||||
// (pkt==NULL means get more output, pkt->size==0 is a flush/drain packet)
|
||||
static int decode(InputStream *ist, AVCodecContext *avctx,
|
||||
AVFrame *frame, int *got_frame, const AVPacket *pkt)
|
||||
{
|
||||
int ret;
|
||||
|
||||
*got_frame = 0;
|
||||
|
||||
if (pkt) {
|
||||
ret = avcodec_send_packet(avctx, pkt);
|
||||
// In particular, we don't expect AVERROR(EAGAIN), because we read all
|
||||
// decoded frames with avcodec_receive_frame() until done.
|
||||
if (ret < 0 && ret != AVERROR_EOF)
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = avcodec_receive_frame(avctx, frame);
|
||||
if (ret < 0 && ret != AVERROR(EAGAIN))
|
||||
return ret;
|
||||
if (ret >= 0) {
|
||||
if (ist->want_frame_data) {
|
||||
FrameData *fd;
|
||||
|
||||
av_assert0(!frame->opaque_ref);
|
||||
frame->opaque_ref = av_buffer_allocz(sizeof(*fd));
|
||||
if (!frame->opaque_ref) {
|
||||
av_frame_unref(frame);
|
||||
return AVERROR(ENOMEM);
|
||||
}
|
||||
fd = (FrameData*)frame->opaque_ref->data;
|
||||
fd->pts = frame->pts;
|
||||
fd->tb = avctx->pkt_timebase;
|
||||
fd->idx = avctx->frame_num - 1;
|
||||
}
|
||||
|
||||
frame->time_base = avctx->pkt_timebase;
|
||||
|
||||
*got_frame = 1;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int send_frame_to_filters(InputStream *ist, AVFrame *decoded_frame)
|
||||
{
|
||||
int i, ret;
|
||||
|
||||
av_assert1(ist->nb_filters > 0); /* ensure ret is initialized */
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
ret = ifilter_send_frame(ist->filters[i], decoded_frame, i < ist->nb_filters - 1);
|
||||
if (ret == AVERROR_EOF)
|
||||
ret = 0; /* ignore */
|
||||
if (ret < 0) {
|
||||
av_log(NULL, AV_LOG_ERROR,
|
||||
"Failed to inject frame into filter network: %s\n", av_err2str(ret));
|
||||
break;
|
||||
}
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
static AVRational audio_samplerate_update(InputStream *ist, const AVFrame *frame)
|
||||
{
|
||||
const int prev = ist->last_frame_tb.den;
|
||||
const int sr = frame->sample_rate;
|
||||
|
||||
AVRational tb_new;
|
||||
int64_t gcd;
|
||||
|
||||
if (frame->sample_rate == ist->last_frame_sample_rate)
|
||||
goto finish;
|
||||
|
||||
gcd = av_gcd(prev, sr);
|
||||
|
||||
if (prev / gcd >= INT_MAX / sr) {
|
||||
av_log(ist, AV_LOG_WARNING,
|
||||
"Audio timestamps cannot be represented exactly after "
|
||||
"sample rate change: %d -> %d\n", prev, sr);
|
||||
|
||||
// LCM of 192000, 44100, allows to represent all common samplerates
|
||||
tb_new = (AVRational){ 1, 28224000 };
|
||||
} else
|
||||
tb_new = (AVRational){ 1, prev / gcd * sr };
|
||||
|
||||
// keep the frame timebase if it is strictly better than
|
||||
// the samplerate-defined one
|
||||
if (frame->time_base.num == 1 && frame->time_base.den > tb_new.den &&
|
||||
!(frame->time_base.den % tb_new.den))
|
||||
tb_new = frame->time_base;
|
||||
|
||||
if (ist->last_frame_pts != AV_NOPTS_VALUE)
|
||||
ist->last_frame_pts = av_rescale_q(ist->last_frame_pts,
|
||||
ist->last_frame_tb, tb_new);
|
||||
ist->last_frame_duration_est = av_rescale_q(ist->last_frame_duration_est,
|
||||
ist->last_frame_tb, tb_new);
|
||||
|
||||
ist->last_frame_tb = tb_new;
|
||||
ist->last_frame_sample_rate = frame->sample_rate;
|
||||
|
||||
finish:
|
||||
return ist->last_frame_tb;
|
||||
}
|
||||
|
||||
static void audio_ts_process(InputStream *ist, AVFrame *frame)
|
||||
{
|
||||
AVRational tb_filter = (AVRational){1, frame->sample_rate};
|
||||
AVRational tb;
|
||||
int64_t pts_pred;
|
||||
|
||||
// on samplerate change, choose a new internal timebase for timestamp
|
||||
// generation that can represent timestamps from all the samplerates
|
||||
// seen so far
|
||||
tb = audio_samplerate_update(ist, frame);
|
||||
pts_pred = ist->last_frame_pts == AV_NOPTS_VALUE ? 0 :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
|
||||
if (frame->pts == AV_NOPTS_VALUE) {
|
||||
frame->pts = pts_pred;
|
||||
frame->time_base = tb;
|
||||
} else if (ist->last_frame_pts != AV_NOPTS_VALUE &&
|
||||
frame->pts > av_rescale_q_rnd(pts_pred, tb, frame->time_base,
|
||||
AV_ROUND_UP)) {
|
||||
// there was a gap in timestamps, reset conversion state
|
||||
ist->filter_in_rescale_delta_last = AV_NOPTS_VALUE;
|
||||
}
|
||||
|
||||
frame->pts = av_rescale_delta(frame->time_base, frame->pts,
|
||||
tb, frame->nb_samples,
|
||||
&ist->filter_in_rescale_delta_last, tb);
|
||||
|
||||
ist->last_frame_pts = frame->pts;
|
||||
ist->last_frame_duration_est = av_rescale_q(frame->nb_samples,
|
||||
tb_filter, tb);
|
||||
|
||||
// finally convert to filtering timebase
|
||||
frame->pts = av_rescale_q(frame->pts, tb, tb_filter);
|
||||
frame->duration = frame->nb_samples;
|
||||
frame->time_base = tb_filter;
|
||||
}
|
||||
|
||||
static int decode_audio(InputStream *ist, const AVPacket *pkt, int *got_output,
|
||||
int *decode_failed)
|
||||
{
|
||||
AVFrame *decoded_frame = ist->decoded_frame;
|
||||
AVCodecContext *avctx = ist->dec_ctx;
|
||||
int ret, err = 0;
|
||||
|
||||
update_benchmark(NULL);
|
||||
ret = decode(ist, avctx, decoded_frame, got_output, pkt);
|
||||
update_benchmark("decode_audio %d.%d", ist->file_index, ist->st->index);
|
||||
if (ret < 0)
|
||||
*decode_failed = 1;
|
||||
|
||||
if (ret != AVERROR_EOF)
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (!*got_output || ret < 0)
|
||||
return ret;
|
||||
|
||||
ist->samples_decoded += decoded_frame->nb_samples;
|
||||
ist->frames_decoded++;
|
||||
|
||||
audio_ts_process(ist, decoded_frame);
|
||||
|
||||
ist->nb_samples = decoded_frame->nb_samples;
|
||||
err = send_frame_to_filters(ist, decoded_frame);
|
||||
|
||||
av_frame_unref(decoded_frame);
|
||||
return err < 0 ? err : ret;
|
||||
}
|
||||
|
||||
static int64_t video_duration_estimate(const InputStream *ist, const AVFrame *frame)
|
||||
{
|
||||
const InputFile *ifile = input_files[ist->file_index];
|
||||
int64_t codec_duration = 0;
|
||||
|
||||
// XXX lavf currently makes up frame durations when they are not provided by
|
||||
// the container. As there is no way to reliably distinguish real container
|
||||
// durations from the fake made-up ones, we use heuristics based on whether
|
||||
// the container has timestamps. Eventually lavf should stop making up
|
||||
// durations, then this should be simplified.
|
||||
|
||||
// prefer frame duration for containers with timestamps
|
||||
if (frame->duration > 0 && (!ifile->format_nots || ist->framerate.num))
|
||||
return frame->duration;
|
||||
|
||||
if (ist->dec_ctx->framerate.den && ist->dec_ctx->framerate.num) {
|
||||
int fields = frame->repeat_pict + 2;
|
||||
AVRational field_rate = av_mul_q(ist->dec_ctx->framerate,
|
||||
(AVRational){ 2, 1 });
|
||||
codec_duration = av_rescale_q(fields, av_inv_q(field_rate),
|
||||
frame->time_base);
|
||||
}
|
||||
|
||||
// prefer codec-layer duration for containers without timestamps
|
||||
if (codec_duration > 0 && ifile->format_nots)
|
||||
return codec_duration;
|
||||
|
||||
// when timestamps are available, repeat last frame's actual duration
|
||||
// (i.e. pts difference between this and last frame)
|
||||
if (frame->pts != AV_NOPTS_VALUE && ist->last_frame_pts != AV_NOPTS_VALUE &&
|
||||
frame->pts > ist->last_frame_pts)
|
||||
return frame->pts - ist->last_frame_pts;
|
||||
|
||||
// try frame/codec duration
|
||||
if (frame->duration > 0)
|
||||
return frame->duration;
|
||||
if (codec_duration > 0)
|
||||
return codec_duration;
|
||||
|
||||
// try average framerate
|
||||
if (ist->st->avg_frame_rate.num && ist->st->avg_frame_rate.den) {
|
||||
int64_t d = av_rescale_q(1, av_inv_q(ist->st->avg_frame_rate),
|
||||
frame->time_base);
|
||||
if (d > 0)
|
||||
return d;
|
||||
}
|
||||
|
||||
// last resort is last frame's estimated duration, and 1
|
||||
return FFMAX(ist->last_frame_duration_est, 1);
|
||||
}
|
||||
|
||||
static int decode_video(InputStream *ist, const AVPacket *pkt, int *got_output,
|
||||
int eof, int *decode_failed)
|
||||
{
|
||||
AVFrame *frame = ist->decoded_frame;
|
||||
int ret = 0, err = 0;
|
||||
|
||||
// With fate-indeo3-2, we're getting 0-sized packets before EOF for some
|
||||
// reason. This seems like a semi-critical bug. Don't trigger EOF, and
|
||||
// skip the packet.
|
||||
if (!eof && pkt && pkt->size == 0)
|
||||
return 0;
|
||||
|
||||
update_benchmark(NULL);
|
||||
ret = decode(ist, ist->dec_ctx, frame, got_output, pkt);
|
||||
update_benchmark("decode_video %d.%d", ist->file_index, ist->st->index);
|
||||
if (ret < 0)
|
||||
*decode_failed = 1;
|
||||
|
||||
// The following line may be required in some cases where there is no parser
|
||||
// or the parser does not has_b_frames correctly
|
||||
if (ist->par->video_delay < ist->dec_ctx->has_b_frames) {
|
||||
if (ist->dec_ctx->codec_id == AV_CODEC_ID_H264) {
|
||||
ist->par->video_delay = ist->dec_ctx->has_b_frames;
|
||||
} else
|
||||
av_log(ist->dec_ctx, AV_LOG_WARNING,
|
||||
"video_delay is larger in decoder than demuxer %d > %d.\n"
|
||||
"If you want to help, upload a sample "
|
||||
"of this file to https://streams.videolan.org/upload/ "
|
||||
"and contact the ffmpeg-devel mailing list. (ffmpeg-devel@ffmpeg.org)\n",
|
||||
ist->dec_ctx->has_b_frames,
|
||||
ist->par->video_delay);
|
||||
}
|
||||
|
||||
if (ret != AVERROR_EOF)
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (*got_output && ret >= 0) {
|
||||
if (ist->dec_ctx->width != frame->width ||
|
||||
ist->dec_ctx->height != frame->height ||
|
||||
ist->dec_ctx->pix_fmt != frame->format) {
|
||||
av_log(NULL, AV_LOG_DEBUG, "Frame parameters mismatch context %d,%d,%d != %d,%d,%d\n",
|
||||
frame->width,
|
||||
frame->height,
|
||||
frame->format,
|
||||
ist->dec_ctx->width,
|
||||
ist->dec_ctx->height,
|
||||
ist->dec_ctx->pix_fmt);
|
||||
}
|
||||
}
|
||||
|
||||
if (!*got_output || ret < 0)
|
||||
return ret;
|
||||
|
||||
if(ist->top_field_first>=0)
|
||||
frame->flags |= AV_FRAME_FLAG_TOP_FIELD_FIRST;
|
||||
|
||||
ist->frames_decoded++;
|
||||
|
||||
if (ist->hwaccel_retrieve_data && frame->format == ist->hwaccel_pix_fmt) {
|
||||
err = ist->hwaccel_retrieve_data(ist->dec_ctx, frame);
|
||||
if (err < 0)
|
||||
goto fail;
|
||||
}
|
||||
|
||||
frame->pts = frame->best_effort_timestamp;
|
||||
|
||||
// forced fixed framerate
|
||||
if (ist->framerate.num) {
|
||||
frame->pts = AV_NOPTS_VALUE;
|
||||
frame->duration = 1;
|
||||
frame->time_base = av_inv_q(ist->framerate);
|
||||
}
|
||||
|
||||
// no timestamp available - extrapolate from previous frame duration
|
||||
if (frame->pts == AV_NOPTS_VALUE)
|
||||
frame->pts = ist->last_frame_pts == AV_NOPTS_VALUE ? 0 :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
|
||||
// update timestamp history
|
||||
ist->last_frame_duration_est = video_duration_estimate(ist, frame);
|
||||
ist->last_frame_pts = frame->pts;
|
||||
ist->last_frame_tb = frame->time_base;
|
||||
|
||||
if (debug_ts) {
|
||||
av_log(ist, AV_LOG_INFO,
|
||||
"decoder -> pts:%s pts_time:%s "
|
||||
"pkt_dts:%s pkt_dts_time:%s "
|
||||
"duration:%s duration_time:%s "
|
||||
"keyframe:%d frame_type:%d time_base:%d/%d\n",
|
||||
av_ts2str(frame->pts),
|
||||
av_ts2timestr(frame->pts, &frame->time_base),
|
||||
av_ts2str(frame->pkt_dts),
|
||||
av_ts2timestr(frame->pkt_dts, &frame->time_base),
|
||||
av_ts2str(frame->duration),
|
||||
av_ts2timestr(frame->duration, &frame->time_base),
|
||||
!!(frame->flags & AV_FRAME_FLAG_KEY), frame->pict_type,
|
||||
frame->time_base.num, frame->time_base.den);
|
||||
}
|
||||
|
||||
if (ist->st->sample_aspect_ratio.num)
|
||||
frame->sample_aspect_ratio = ist->st->sample_aspect_ratio;
|
||||
|
||||
err = send_frame_to_filters(ist, frame);
|
||||
|
||||
fail:
|
||||
av_frame_unref(frame);
|
||||
return err < 0 ? err : ret;
|
||||
}
|
||||
|
||||
static int process_subtitle(InputStream *ist, AVSubtitle *subtitle, int *got_output)
|
||||
{
|
||||
int ret = 0;
|
||||
int free_sub = 1;
|
||||
|
||||
if (ist->fix_sub_duration) {
|
||||
int end = 1;
|
||||
if (ist->prev_sub.got_output) {
|
||||
end = av_rescale(subtitle->pts - ist->prev_sub.subtitle.pts,
|
||||
1000, AV_TIME_BASE);
|
||||
if (end < ist->prev_sub.subtitle.end_display_time) {
|
||||
av_log(NULL, AV_LOG_DEBUG,
|
||||
"Subtitle duration reduced from %"PRId32" to %d%s\n",
|
||||
ist->prev_sub.subtitle.end_display_time, end,
|
||||
end <= 0 ? ", dropping it" : "");
|
||||
ist->prev_sub.subtitle.end_display_time = end;
|
||||
}
|
||||
}
|
||||
FFSWAP(int, *got_output, ist->prev_sub.got_output);
|
||||
FFSWAP(int, ret, ist->prev_sub.ret);
|
||||
FFSWAP(AVSubtitle, *subtitle, ist->prev_sub.subtitle);
|
||||
if (end <= 0)
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (!*got_output)
|
||||
return ret;
|
||||
|
||||
if (ist->sub2video.frame) {
|
||||
sub2video_update(ist, INT64_MIN, subtitle);
|
||||
} else if (ist->nb_filters) {
|
||||
if (!ist->sub2video.sub_queue)
|
||||
ist->sub2video.sub_queue = av_fifo_alloc2(8, sizeof(AVSubtitle), AV_FIFO_FLAG_AUTO_GROW);
|
||||
if (!ist->sub2video.sub_queue)
|
||||
report_and_exit(AVERROR(ENOMEM));
|
||||
|
||||
ret = av_fifo_write(ist->sub2video.sub_queue, subtitle, 1);
|
||||
if (ret < 0)
|
||||
exit_program(1);
|
||||
free_sub = 0;
|
||||
}
|
||||
|
||||
if (!subtitle->num_rects)
|
||||
goto out;
|
||||
|
||||
for (int oidx = 0; oidx < ist->nb_outputs; oidx++) {
|
||||
OutputStream *ost = ist->outputs[oidx];
|
||||
if (!ost->enc || ost->type != AVMEDIA_TYPE_SUBTITLE)
|
||||
continue;
|
||||
|
||||
enc_subtitle(output_files[ost->file_index], ost, subtitle);
|
||||
}
|
||||
|
||||
out:
|
||||
if (free_sub)
|
||||
avsubtitle_free(subtitle);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int copy_av_subtitle(AVSubtitle *dst, AVSubtitle *src)
|
||||
{
|
||||
int ret = AVERROR_BUG;
|
||||
@ -1309,115 +884,6 @@ int trigger_fix_sub_duration_heartbeat(OutputStream *ost, const AVPacket *pkt)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int transcode_subtitles(InputStream *ist, const AVPacket *pkt,
|
||||
int *got_output, int *decode_failed)
|
||||
{
|
||||
AVSubtitle subtitle;
|
||||
int ret = avcodec_decode_subtitle2(ist->dec_ctx,
|
||||
&subtitle, got_output, pkt);
|
||||
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (ret < 0 || !*got_output) {
|
||||
*decode_failed = 1;
|
||||
if (!pkt->size)
|
||||
sub2video_flush(ist);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ist->frames_decoded++;
|
||||
|
||||
return process_subtitle(ist, &subtitle, got_output);
|
||||
}
|
||||
|
||||
static int send_filter_eof(InputStream *ist)
|
||||
{
|
||||
int i, ret;
|
||||
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
int64_t end_pts = ist->last_frame_pts == AV_NOPTS_VALUE ? AV_NOPTS_VALUE :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
ret = ifilter_send_eof(ist->filters[i], end_pts, ist->last_frame_tb);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int dec_packet(InputStream *ist, const AVPacket *pkt, int no_eof)
|
||||
{
|
||||
AVPacket *avpkt = ist->pkt;
|
||||
int ret, repeating = 0;
|
||||
|
||||
if (pkt) {
|
||||
av_packet_unref(avpkt);
|
||||
ret = av_packet_ref(avpkt, pkt);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
}
|
||||
|
||||
// while we have more to decode or while the decoder did output something on EOF
|
||||
while (1) {
|
||||
int got_output = 0;
|
||||
int decode_failed = 0;
|
||||
|
||||
switch (ist->par->codec_type) {
|
||||
case AVMEDIA_TYPE_AUDIO:
|
||||
ret = decode_audio (ist, repeating ? NULL : avpkt, &got_output,
|
||||
&decode_failed);
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
case AVMEDIA_TYPE_VIDEO:
|
||||
ret = decode_video (ist, repeating ? NULL : avpkt, &got_output, !pkt,
|
||||
&decode_failed);
|
||||
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
case AVMEDIA_TYPE_SUBTITLE:
|
||||
if (repeating)
|
||||
break;
|
||||
ret = transcode_subtitles(ist, avpkt, &got_output, &decode_failed);
|
||||
if (!pkt && ret >= 0)
|
||||
ret = AVERROR_EOF;
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
default: av_assert0(0);
|
||||
}
|
||||
|
||||
if (ret == AVERROR_EOF) {
|
||||
/* after flushing, send an EOF on all the filter inputs attached to the stream */
|
||||
/* except when looping we need to flush but not to send an EOF */
|
||||
if (!no_eof) {
|
||||
ret = send_filter_eof(ist);
|
||||
if (ret < 0) {
|
||||
av_log(NULL, AV_LOG_FATAL, "Error marking filters as finished\n");
|
||||
exit_program(1);
|
||||
}
|
||||
}
|
||||
|
||||
return AVERROR_EOF;
|
||||
}
|
||||
|
||||
if (ret < 0) {
|
||||
if (decode_failed) {
|
||||
av_log(NULL, AV_LOG_ERROR, "Error while decoding stream #%d:%d: %s\n",
|
||||
ist->file_index, ist->st->index, av_err2str(ret));
|
||||
} else {
|
||||
av_log(NULL, AV_LOG_FATAL, "Error while processing the decoded "
|
||||
"data for stream #%d:%d\n", ist->file_index, ist->st->index);
|
||||
}
|
||||
if (!decode_failed || exit_on_error)
|
||||
exit_program(1);
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (!got_output)
|
||||
return 0;
|
||||
|
||||
repeating = 1;
|
||||
}
|
||||
}
|
||||
|
||||
/* pkt = NULL means EOF (needed to flush decoder buffers) */
|
||||
static int process_input_packet(InputStream *ist, const AVPacket *pkt, int no_eof)
|
||||
{
|
||||
|
@ -811,6 +811,17 @@ int hwaccel_decode_init(AVCodecContext *avctx);
|
||||
|
||||
int dec_open(InputStream *ist);
|
||||
|
||||
/**
|
||||
* Submit a packet for decoding
|
||||
*
|
||||
* When pkt==NULL and no_eof=0, there will be no more input. Flush decoders and
|
||||
* mark all downstreams as finished.
|
||||
*
|
||||
* When pkt==NULL and no_eof=1, the stream was reset (e.g. after a seek). Flush
|
||||
* decoders and await further input.
|
||||
*/
|
||||
int dec_packet(InputStream *ist, const AVPacket *pkt, int no_eof);
|
||||
|
||||
int enc_alloc(Encoder **penc, const AVCodec *codec);
|
||||
void enc_free(Encoder **penc);
|
||||
|
||||
@ -885,6 +896,7 @@ OutputStream *ost_iter(OutputStream *prev);
|
||||
|
||||
void close_output_stream(OutputStream *ost);
|
||||
int trigger_fix_sub_duration_heartbeat(OutputStream *ost, const AVPacket *pkt);
|
||||
int process_subtitle(InputStream *ist, AVSubtitle *subtitle, int *got_output);
|
||||
void update_benchmark(const char *fmt, ...);
|
||||
|
||||
/**
|
||||
|
@ -16,17 +16,555 @@
|
||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
||||
*/
|
||||
|
||||
#include "libavutil/avassert.h"
|
||||
#include "libavutil/dict.h"
|
||||
#include "libavutil/error.h"
|
||||
#include "libavutil/log.h"
|
||||
#include "libavutil/pixdesc.h"
|
||||
#include "libavutil/pixfmt.h"
|
||||
#include "libavutil/timestamp.h"
|
||||
|
||||
#include "libavcodec/avcodec.h"
|
||||
#include "libavcodec/codec.h"
|
||||
|
||||
#include "libavfilter/buffersrc.h"
|
||||
|
||||
#include "ffmpeg.h"
|
||||
|
||||
static void check_decode_result(InputStream *ist, int *got_output, int ret)
|
||||
{
|
||||
if (ret < 0)
|
||||
ist->decode_errors++;
|
||||
|
||||
if (ret < 0 && exit_on_error)
|
||||
exit_program(1);
|
||||
|
||||
if (*got_output && ist->dec_ctx->codec_type != AVMEDIA_TYPE_SUBTITLE) {
|
||||
if (ist->decoded_frame->decode_error_flags || (ist->decoded_frame->flags & AV_FRAME_FLAG_CORRUPT)) {
|
||||
av_log(ist, exit_on_error ? AV_LOG_FATAL : AV_LOG_WARNING,
|
||||
"corrupt decoded frame\n");
|
||||
if (exit_on_error)
|
||||
exit_program(1);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// This does not quite work like avcodec_decode_audio4/avcodec_decode_video2.
|
||||
// There is the following difference: if you got a frame, you must call
|
||||
// it again with pkt=NULL. pkt==NULL is treated differently from pkt->size==0
|
||||
// (pkt==NULL means get more output, pkt->size==0 is a flush/drain packet)
|
||||
static int decode(InputStream *ist, AVCodecContext *avctx,
|
||||
AVFrame *frame, int *got_frame, const AVPacket *pkt)
|
||||
{
|
||||
int ret;
|
||||
|
||||
*got_frame = 0;
|
||||
|
||||
if (pkt) {
|
||||
ret = avcodec_send_packet(avctx, pkt);
|
||||
// In particular, we don't expect AVERROR(EAGAIN), because we read all
|
||||
// decoded frames with avcodec_receive_frame() until done.
|
||||
if (ret < 0 && ret != AVERROR_EOF)
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = avcodec_receive_frame(avctx, frame);
|
||||
if (ret < 0 && ret != AVERROR(EAGAIN))
|
||||
return ret;
|
||||
if (ret >= 0) {
|
||||
if (ist->want_frame_data) {
|
||||
FrameData *fd;
|
||||
|
||||
av_assert0(!frame->opaque_ref);
|
||||
frame->opaque_ref = av_buffer_allocz(sizeof(*fd));
|
||||
if (!frame->opaque_ref) {
|
||||
av_frame_unref(frame);
|
||||
return AVERROR(ENOMEM);
|
||||
}
|
||||
fd = (FrameData*)frame->opaque_ref->data;
|
||||
fd->pts = frame->pts;
|
||||
fd->tb = avctx->pkt_timebase;
|
||||
fd->idx = avctx->frame_num - 1;
|
||||
}
|
||||
|
||||
frame->time_base = avctx->pkt_timebase;
|
||||
|
||||
*got_frame = 1;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int send_frame_to_filters(InputStream *ist, AVFrame *decoded_frame)
|
||||
{
|
||||
int i, ret;
|
||||
|
||||
av_assert1(ist->nb_filters > 0); /* ensure ret is initialized */
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
ret = ifilter_send_frame(ist->filters[i], decoded_frame, i < ist->nb_filters - 1);
|
||||
if (ret == AVERROR_EOF)
|
||||
ret = 0; /* ignore */
|
||||
if (ret < 0) {
|
||||
av_log(NULL, AV_LOG_ERROR,
|
||||
"Failed to inject frame into filter network: %s\n", av_err2str(ret));
|
||||
break;
|
||||
}
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
static AVRational audio_samplerate_update(InputStream *ist, const AVFrame *frame)
|
||||
{
|
||||
const int prev = ist->last_frame_tb.den;
|
||||
const int sr = frame->sample_rate;
|
||||
|
||||
AVRational tb_new;
|
||||
int64_t gcd;
|
||||
|
||||
if (frame->sample_rate == ist->last_frame_sample_rate)
|
||||
goto finish;
|
||||
|
||||
gcd = av_gcd(prev, sr);
|
||||
|
||||
if (prev / gcd >= INT_MAX / sr) {
|
||||
av_log(ist, AV_LOG_WARNING,
|
||||
"Audio timestamps cannot be represented exactly after "
|
||||
"sample rate change: %d -> %d\n", prev, sr);
|
||||
|
||||
// LCM of 192000, 44100, allows to represent all common samplerates
|
||||
tb_new = (AVRational){ 1, 28224000 };
|
||||
} else
|
||||
tb_new = (AVRational){ 1, prev / gcd * sr };
|
||||
|
||||
// keep the frame timebase if it is strictly better than
|
||||
// the samplerate-defined one
|
||||
if (frame->time_base.num == 1 && frame->time_base.den > tb_new.den &&
|
||||
!(frame->time_base.den % tb_new.den))
|
||||
tb_new = frame->time_base;
|
||||
|
||||
if (ist->last_frame_pts != AV_NOPTS_VALUE)
|
||||
ist->last_frame_pts = av_rescale_q(ist->last_frame_pts,
|
||||
ist->last_frame_tb, tb_new);
|
||||
ist->last_frame_duration_est = av_rescale_q(ist->last_frame_duration_est,
|
||||
ist->last_frame_tb, tb_new);
|
||||
|
||||
ist->last_frame_tb = tb_new;
|
||||
ist->last_frame_sample_rate = frame->sample_rate;
|
||||
|
||||
finish:
|
||||
return ist->last_frame_tb;
|
||||
}
|
||||
|
||||
static void audio_ts_process(InputStream *ist, AVFrame *frame)
|
||||
{
|
||||
AVRational tb_filter = (AVRational){1, frame->sample_rate};
|
||||
AVRational tb;
|
||||
int64_t pts_pred;
|
||||
|
||||
// on samplerate change, choose a new internal timebase for timestamp
|
||||
// generation that can represent timestamps from all the samplerates
|
||||
// seen so far
|
||||
tb = audio_samplerate_update(ist, frame);
|
||||
pts_pred = ist->last_frame_pts == AV_NOPTS_VALUE ? 0 :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
|
||||
if (frame->pts == AV_NOPTS_VALUE) {
|
||||
frame->pts = pts_pred;
|
||||
frame->time_base = tb;
|
||||
} else if (ist->last_frame_pts != AV_NOPTS_VALUE &&
|
||||
frame->pts > av_rescale_q_rnd(pts_pred, tb, frame->time_base,
|
||||
AV_ROUND_UP)) {
|
||||
// there was a gap in timestamps, reset conversion state
|
||||
ist->filter_in_rescale_delta_last = AV_NOPTS_VALUE;
|
||||
}
|
||||
|
||||
frame->pts = av_rescale_delta(frame->time_base, frame->pts,
|
||||
tb, frame->nb_samples,
|
||||
&ist->filter_in_rescale_delta_last, tb);
|
||||
|
||||
ist->last_frame_pts = frame->pts;
|
||||
ist->last_frame_duration_est = av_rescale_q(frame->nb_samples,
|
||||
tb_filter, tb);
|
||||
|
||||
// finally convert to filtering timebase
|
||||
frame->pts = av_rescale_q(frame->pts, tb, tb_filter);
|
||||
frame->duration = frame->nb_samples;
|
||||
frame->time_base = tb_filter;
|
||||
}
|
||||
|
||||
static int decode_audio(InputStream *ist, const AVPacket *pkt, int *got_output,
|
||||
int *decode_failed)
|
||||
{
|
||||
AVFrame *decoded_frame = ist->decoded_frame;
|
||||
AVCodecContext *avctx = ist->dec_ctx;
|
||||
int ret, err = 0;
|
||||
|
||||
update_benchmark(NULL);
|
||||
ret = decode(ist, avctx, decoded_frame, got_output, pkt);
|
||||
update_benchmark("decode_audio %d.%d", ist->file_index, ist->st->index);
|
||||
if (ret < 0)
|
||||
*decode_failed = 1;
|
||||
|
||||
if (ret != AVERROR_EOF)
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (!*got_output || ret < 0)
|
||||
return ret;
|
||||
|
||||
ist->samples_decoded += decoded_frame->nb_samples;
|
||||
ist->frames_decoded++;
|
||||
|
||||
audio_ts_process(ist, decoded_frame);
|
||||
|
||||
ist->nb_samples = decoded_frame->nb_samples;
|
||||
err = send_frame_to_filters(ist, decoded_frame);
|
||||
|
||||
av_frame_unref(decoded_frame);
|
||||
return err < 0 ? err : ret;
|
||||
}
|
||||
|
||||
static int64_t video_duration_estimate(const InputStream *ist, const AVFrame *frame)
|
||||
{
|
||||
const InputFile *ifile = input_files[ist->file_index];
|
||||
int64_t codec_duration = 0;
|
||||
|
||||
// XXX lavf currently makes up frame durations when they are not provided by
|
||||
// the container. As there is no way to reliably distinguish real container
|
||||
// durations from the fake made-up ones, we use heuristics based on whether
|
||||
// the container has timestamps. Eventually lavf should stop making up
|
||||
// durations, then this should be simplified.
|
||||
|
||||
// prefer frame duration for containers with timestamps
|
||||
if (frame->duration > 0 && (!ifile->format_nots || ist->framerate.num))
|
||||
return frame->duration;
|
||||
|
||||
if (ist->dec_ctx->framerate.den && ist->dec_ctx->framerate.num) {
|
||||
int fields = frame->repeat_pict + 2;
|
||||
AVRational field_rate = av_mul_q(ist->dec_ctx->framerate,
|
||||
(AVRational){ 2, 1 });
|
||||
codec_duration = av_rescale_q(fields, av_inv_q(field_rate),
|
||||
frame->time_base);
|
||||
}
|
||||
|
||||
// prefer codec-layer duration for containers without timestamps
|
||||
if (codec_duration > 0 && ifile->format_nots)
|
||||
return codec_duration;
|
||||
|
||||
// when timestamps are available, repeat last frame's actual duration
|
||||
// (i.e. pts difference between this and last frame)
|
||||
if (frame->pts != AV_NOPTS_VALUE && ist->last_frame_pts != AV_NOPTS_VALUE &&
|
||||
frame->pts > ist->last_frame_pts)
|
||||
return frame->pts - ist->last_frame_pts;
|
||||
|
||||
// try frame/codec duration
|
||||
if (frame->duration > 0)
|
||||
return frame->duration;
|
||||
if (codec_duration > 0)
|
||||
return codec_duration;
|
||||
|
||||
// try average framerate
|
||||
if (ist->st->avg_frame_rate.num && ist->st->avg_frame_rate.den) {
|
||||
int64_t d = av_rescale_q(1, av_inv_q(ist->st->avg_frame_rate),
|
||||
frame->time_base);
|
||||
if (d > 0)
|
||||
return d;
|
||||
}
|
||||
|
||||
// last resort is last frame's estimated duration, and 1
|
||||
return FFMAX(ist->last_frame_duration_est, 1);
|
||||
}
|
||||
|
||||
static int decode_video(InputStream *ist, const AVPacket *pkt, int *got_output,
|
||||
int eof, int *decode_failed)
|
||||
{
|
||||
AVFrame *frame = ist->decoded_frame;
|
||||
int ret = 0, err = 0;
|
||||
|
||||
// With fate-indeo3-2, we're getting 0-sized packets before EOF for some
|
||||
// reason. This seems like a semi-critical bug. Don't trigger EOF, and
|
||||
// skip the packet.
|
||||
if (!eof && pkt && pkt->size == 0)
|
||||
return 0;
|
||||
|
||||
update_benchmark(NULL);
|
||||
ret = decode(ist, ist->dec_ctx, frame, got_output, pkt);
|
||||
update_benchmark("decode_video %d.%d", ist->file_index, ist->st->index);
|
||||
if (ret < 0)
|
||||
*decode_failed = 1;
|
||||
|
||||
// The following line may be required in some cases where there is no parser
|
||||
// or the parser does not has_b_frames correctly
|
||||
if (ist->par->video_delay < ist->dec_ctx->has_b_frames) {
|
||||
if (ist->dec_ctx->codec_id == AV_CODEC_ID_H264) {
|
||||
ist->par->video_delay = ist->dec_ctx->has_b_frames;
|
||||
} else
|
||||
av_log(ist->dec_ctx, AV_LOG_WARNING,
|
||||
"video_delay is larger in decoder than demuxer %d > %d.\n"
|
||||
"If you want to help, upload a sample "
|
||||
"of this file to https://streams.videolan.org/upload/ "
|
||||
"and contact the ffmpeg-devel mailing list. (ffmpeg-devel@ffmpeg.org)\n",
|
||||
ist->dec_ctx->has_b_frames,
|
||||
ist->par->video_delay);
|
||||
}
|
||||
|
||||
if (ret != AVERROR_EOF)
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (*got_output && ret >= 0) {
|
||||
if (ist->dec_ctx->width != frame->width ||
|
||||
ist->dec_ctx->height != frame->height ||
|
||||
ist->dec_ctx->pix_fmt != frame->format) {
|
||||
av_log(NULL, AV_LOG_DEBUG, "Frame parameters mismatch context %d,%d,%d != %d,%d,%d\n",
|
||||
frame->width,
|
||||
frame->height,
|
||||
frame->format,
|
||||
ist->dec_ctx->width,
|
||||
ist->dec_ctx->height,
|
||||
ist->dec_ctx->pix_fmt);
|
||||
}
|
||||
}
|
||||
|
||||
if (!*got_output || ret < 0)
|
||||
return ret;
|
||||
|
||||
if(ist->top_field_first>=0)
|
||||
frame->flags |= AV_FRAME_FLAG_TOP_FIELD_FIRST;
|
||||
|
||||
ist->frames_decoded++;
|
||||
|
||||
if (ist->hwaccel_retrieve_data && frame->format == ist->hwaccel_pix_fmt) {
|
||||
err = ist->hwaccel_retrieve_data(ist->dec_ctx, frame);
|
||||
if (err < 0)
|
||||
goto fail;
|
||||
}
|
||||
|
||||
frame->pts = frame->best_effort_timestamp;
|
||||
|
||||
// forced fixed framerate
|
||||
if (ist->framerate.num) {
|
||||
frame->pts = AV_NOPTS_VALUE;
|
||||
frame->duration = 1;
|
||||
frame->time_base = av_inv_q(ist->framerate);
|
||||
}
|
||||
|
||||
// no timestamp available - extrapolate from previous frame duration
|
||||
if (frame->pts == AV_NOPTS_VALUE)
|
||||
frame->pts = ist->last_frame_pts == AV_NOPTS_VALUE ? 0 :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
|
||||
// update timestamp history
|
||||
ist->last_frame_duration_est = video_duration_estimate(ist, frame);
|
||||
ist->last_frame_pts = frame->pts;
|
||||
ist->last_frame_tb = frame->time_base;
|
||||
|
||||
if (debug_ts) {
|
||||
av_log(ist, AV_LOG_INFO,
|
||||
"decoder -> pts:%s pts_time:%s "
|
||||
"pkt_dts:%s pkt_dts_time:%s "
|
||||
"duration:%s duration_time:%s "
|
||||
"keyframe:%d frame_type:%d time_base:%d/%d\n",
|
||||
av_ts2str(frame->pts),
|
||||
av_ts2timestr(frame->pts, &frame->time_base),
|
||||
av_ts2str(frame->pkt_dts),
|
||||
av_ts2timestr(frame->pkt_dts, &frame->time_base),
|
||||
av_ts2str(frame->duration),
|
||||
av_ts2timestr(frame->duration, &frame->time_base),
|
||||
!!(frame->flags & AV_FRAME_FLAG_KEY), frame->pict_type,
|
||||
frame->time_base.num, frame->time_base.den);
|
||||
}
|
||||
|
||||
if (ist->st->sample_aspect_ratio.num)
|
||||
frame->sample_aspect_ratio = ist->st->sample_aspect_ratio;
|
||||
|
||||
err = send_frame_to_filters(ist, frame);
|
||||
|
||||
fail:
|
||||
av_frame_unref(frame);
|
||||
return err < 0 ? err : ret;
|
||||
}
|
||||
|
||||
static void sub2video_flush(InputStream *ist)
|
||||
{
|
||||
int i;
|
||||
int ret;
|
||||
|
||||
if (ist->sub2video.end_pts < INT64_MAX)
|
||||
sub2video_update(ist, INT64_MAX, NULL);
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
ret = av_buffersrc_add_frame(ist->filters[i]->filter, NULL);
|
||||
if (ret != AVERROR_EOF && ret < 0)
|
||||
av_log(NULL, AV_LOG_WARNING, "Flush the frame error.\n");
|
||||
}
|
||||
}
|
||||
|
||||
int process_subtitle(InputStream *ist, AVSubtitle *subtitle, int *got_output)
|
||||
{
|
||||
int ret = 0;
|
||||
int free_sub = 1;
|
||||
|
||||
if (ist->fix_sub_duration) {
|
||||
int end = 1;
|
||||
if (ist->prev_sub.got_output) {
|
||||
end = av_rescale(subtitle->pts - ist->prev_sub.subtitle.pts,
|
||||
1000, AV_TIME_BASE);
|
||||
if (end < ist->prev_sub.subtitle.end_display_time) {
|
||||
av_log(NULL, AV_LOG_DEBUG,
|
||||
"Subtitle duration reduced from %"PRId32" to %d%s\n",
|
||||
ist->prev_sub.subtitle.end_display_time, end,
|
||||
end <= 0 ? ", dropping it" : "");
|
||||
ist->prev_sub.subtitle.end_display_time = end;
|
||||
}
|
||||
}
|
||||
FFSWAP(int, *got_output, ist->prev_sub.got_output);
|
||||
FFSWAP(int, ret, ist->prev_sub.ret);
|
||||
FFSWAP(AVSubtitle, *subtitle, ist->prev_sub.subtitle);
|
||||
if (end <= 0)
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (!*got_output)
|
||||
return ret;
|
||||
|
||||
if (ist->sub2video.frame) {
|
||||
sub2video_update(ist, INT64_MIN, subtitle);
|
||||
} else if (ist->nb_filters) {
|
||||
if (!ist->sub2video.sub_queue)
|
||||
ist->sub2video.sub_queue = av_fifo_alloc2(8, sizeof(AVSubtitle), AV_FIFO_FLAG_AUTO_GROW);
|
||||
if (!ist->sub2video.sub_queue)
|
||||
report_and_exit(AVERROR(ENOMEM));
|
||||
|
||||
ret = av_fifo_write(ist->sub2video.sub_queue, subtitle, 1);
|
||||
if (ret < 0)
|
||||
exit_program(1);
|
||||
free_sub = 0;
|
||||
}
|
||||
|
||||
if (!subtitle->num_rects)
|
||||
goto out;
|
||||
|
||||
for (int oidx = 0; oidx < ist->nb_outputs; oidx++) {
|
||||
OutputStream *ost = ist->outputs[oidx];
|
||||
if (!ost->enc || ost->type != AVMEDIA_TYPE_SUBTITLE)
|
||||
continue;
|
||||
|
||||
enc_subtitle(output_files[ost->file_index], ost, subtitle);
|
||||
}
|
||||
|
||||
out:
|
||||
if (free_sub)
|
||||
avsubtitle_free(subtitle);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int transcode_subtitles(InputStream *ist, const AVPacket *pkt,
|
||||
int *got_output, int *decode_failed)
|
||||
{
|
||||
AVSubtitle subtitle;
|
||||
int ret = avcodec_decode_subtitle2(ist->dec_ctx,
|
||||
&subtitle, got_output, pkt);
|
||||
|
||||
check_decode_result(ist, got_output, ret);
|
||||
|
||||
if (ret < 0 || !*got_output) {
|
||||
*decode_failed = 1;
|
||||
if (!pkt->size)
|
||||
sub2video_flush(ist);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ist->frames_decoded++;
|
||||
|
||||
return process_subtitle(ist, &subtitle, got_output);
|
||||
}
|
||||
|
||||
static int send_filter_eof(InputStream *ist)
|
||||
{
|
||||
int i, ret;
|
||||
|
||||
for (i = 0; i < ist->nb_filters; i++) {
|
||||
int64_t end_pts = ist->last_frame_pts == AV_NOPTS_VALUE ? AV_NOPTS_VALUE :
|
||||
ist->last_frame_pts + ist->last_frame_duration_est;
|
||||
ret = ifilter_send_eof(ist->filters[i], end_pts, ist->last_frame_tb);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
int dec_packet(InputStream *ist, const AVPacket *pkt, int no_eof)
|
||||
{
|
||||
AVPacket *avpkt = ist->pkt;
|
||||
int ret, repeating = 0;
|
||||
|
||||
if (pkt) {
|
||||
av_packet_unref(avpkt);
|
||||
ret = av_packet_ref(avpkt, pkt);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
}
|
||||
|
||||
// while we have more to decode or while the decoder did output something on EOF
|
||||
while (1) {
|
||||
int got_output = 0;
|
||||
int decode_failed = 0;
|
||||
|
||||
switch (ist->par->codec_type) {
|
||||
case AVMEDIA_TYPE_AUDIO:
|
||||
ret = decode_audio (ist, repeating ? NULL : avpkt, &got_output,
|
||||
&decode_failed);
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
case AVMEDIA_TYPE_VIDEO:
|
||||
ret = decode_video (ist, repeating ? NULL : avpkt, &got_output, !pkt,
|
||||
&decode_failed);
|
||||
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
case AVMEDIA_TYPE_SUBTITLE:
|
||||
if (repeating)
|
||||
break;
|
||||
ret = transcode_subtitles(ist, avpkt, &got_output, &decode_failed);
|
||||
if (!pkt && ret >= 0)
|
||||
ret = AVERROR_EOF;
|
||||
av_packet_unref(avpkt);
|
||||
break;
|
||||
default: av_assert0(0);
|
||||
}
|
||||
|
||||
if (ret == AVERROR_EOF) {
|
||||
/* after flushing, send an EOF on all the filter inputs attached to the stream */
|
||||
/* except when looping we need to flush but not to send an EOF */
|
||||
if (!no_eof) {
|
||||
ret = send_filter_eof(ist);
|
||||
if (ret < 0) {
|
||||
av_log(NULL, AV_LOG_FATAL, "Error marking filters as finished\n");
|
||||
exit_program(1);
|
||||
}
|
||||
}
|
||||
|
||||
return AVERROR_EOF;
|
||||
}
|
||||
|
||||
if (ret < 0) {
|
||||
if (decode_failed) {
|
||||
av_log(NULL, AV_LOG_ERROR, "Error while decoding stream #%d:%d: %s\n",
|
||||
ist->file_index, ist->st->index, av_err2str(ret));
|
||||
} else {
|
||||
av_log(NULL, AV_LOG_FATAL, "Error while processing the decoded "
|
||||
"data for stream #%d:%d\n", ist->file_index, ist->st->index);
|
||||
}
|
||||
if (!decode_failed || exit_on_error)
|
||||
exit_program(1);
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (!got_output)
|
||||
return 0;
|
||||
|
||||
repeating = 1;
|
||||
}
|
||||
}
|
||||
|
||||
static enum AVPixelFormat get_format(AVCodecContext *s, const enum AVPixelFormat *pix_fmts)
|
||||
{
|
||||
InputStream *ist = s->opaque;
|
||||
|
Loading…
Reference in New Issue
Block a user