You've already forked FFmpeg
mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2025-08-10 06:10:52 +02:00
fftools/ffmpeg: rework setting sub2video parameters
Set them in ifilter_parameters_from_dec(), similarly to audio/video streams. This reduces the extent to which sub2video filters need to be treated specially.
This commit is contained in:
@@ -153,9 +153,9 @@ static int sub2video_get_blank_frame(InputStream *ist)
|
|||||||
AVFrame *frame = ist->sub2video.frame;
|
AVFrame *frame = ist->sub2video.frame;
|
||||||
|
|
||||||
av_frame_unref(frame);
|
av_frame_unref(frame);
|
||||||
ist->sub2video.frame->width = ist->dec_ctx->width ? ist->dec_ctx->width : ist->sub2video.w;
|
frame->width = ist->sub2video.w;
|
||||||
ist->sub2video.frame->height = ist->dec_ctx->height ? ist->dec_ctx->height : ist->sub2video.h;
|
frame->height = ist->sub2video.h;
|
||||||
ist->sub2video.frame->format = AV_PIX_FMT_RGB32;
|
frame->format = AV_PIX_FMT_RGB32;
|
||||||
if ((ret = av_frame_get_buffer(frame, 0)) < 0)
|
if ((ret = av_frame_get_buffer(frame, 0)) < 0)
|
||||||
return ret;
|
return ret;
|
||||||
memset(frame->data[0], 0, frame->height * frame->linesize[0]);
|
memset(frame->data[0], 0, frame->height * frame->linesize[0]);
|
||||||
|
@@ -1228,6 +1228,27 @@ static void add_input_streams(const OptionsContext *o, Demuxer *d)
|
|||||||
av_log(ist, AV_LOG_FATAL, "Invalid canvas size: %s.\n", canvas_size);
|
av_log(ist, AV_LOG_FATAL, "Invalid canvas size: %s.\n", canvas_size);
|
||||||
exit_program(1);
|
exit_program(1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Compute the size of the canvas for the subtitles stream.
|
||||||
|
If the subtitles codecpar has set a size, use it. Otherwise use the
|
||||||
|
maximum dimensions of the video streams in the same file. */
|
||||||
|
ist->sub2video.w = ist->dec_ctx->width;
|
||||||
|
ist->sub2video.h = ist->dec_ctx->height;
|
||||||
|
if (!(ist->sub2video.w && ist->sub2video.h)) {
|
||||||
|
for (int j = 0; j < ic->nb_streams; j++) {
|
||||||
|
AVCodecParameters *par1 = ic->streams[j]->codecpar;
|
||||||
|
if (par1->codec_type == AVMEDIA_TYPE_VIDEO) {
|
||||||
|
ist->sub2video.w = FFMAX(ist->sub2video.w, par1->width);
|
||||||
|
ist->sub2video.h = FFMAX(ist->sub2video.h, par1->height);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!(ist->sub2video.w && ist->sub2video.h)) {
|
||||||
|
ist->sub2video.w = FFMAX(ist->sub2video.w, 720);
|
||||||
|
ist->sub2video.h = FFMAX(ist->sub2video.h, 576);
|
||||||
|
}
|
||||||
|
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case AVMEDIA_TYPE_ATTACHMENT:
|
case AVMEDIA_TYPE_ATTACHMENT:
|
||||||
|
@@ -458,13 +458,13 @@ static int ifilter_bind_ist(InputFilter *ifilter, InputStream *ist)
|
|||||||
InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
|
InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
ifp->ist = ist;
|
||||||
|
ifp->type_src = ist->st->codecpar->codec_type;
|
||||||
|
|
||||||
ret = ist_filter_add(ist, ifilter, filtergraph_is_simple(ifilter->graph));
|
ret = ist_filter_add(ist, ifilter, filtergraph_is_simple(ifilter->graph));
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
ifp->ist = ist;
|
|
||||||
ifp->type_src = ist->st->codecpar->codec_type;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1110,38 +1110,6 @@ void check_filter_outputs(void)
|
|||||||
|
|
||||||
static int sub2video_prepare(InputStream *ist, InputFilter *ifilter)
|
static int sub2video_prepare(InputStream *ist, InputFilter *ifilter)
|
||||||
{
|
{
|
||||||
InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
|
|
||||||
AVFormatContext *avf = input_files[ist->file_index]->ctx;
|
|
||||||
int i, w, h;
|
|
||||||
|
|
||||||
/* Compute the size of the canvas for the subtitles stream.
|
|
||||||
If the subtitles codecpar has set a size, use it. Otherwise use the
|
|
||||||
maximum dimensions of the video streams in the same file. */
|
|
||||||
w = ifp->width;
|
|
||||||
h = ifp->height;
|
|
||||||
if (!(w && h)) {
|
|
||||||
for (i = 0; i < avf->nb_streams; i++) {
|
|
||||||
if (avf->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
|
|
||||||
w = FFMAX(w, avf->streams[i]->codecpar->width);
|
|
||||||
h = FFMAX(h, avf->streams[i]->codecpar->height);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (!(w && h)) {
|
|
||||||
w = FFMAX(w, 720);
|
|
||||||
h = FFMAX(h, 576);
|
|
||||||
}
|
|
||||||
av_log(avf, AV_LOG_INFO, "sub2video: using %dx%d canvas\n", w, h);
|
|
||||||
}
|
|
||||||
ist->sub2video.w = ifp->width = w;
|
|
||||||
ist->sub2video.h = ifp->height = h;
|
|
||||||
|
|
||||||
ifp->width = ist->dec_ctx->width ? ist->dec_ctx->width : ist->sub2video.w;
|
|
||||||
ifp->height = ist->dec_ctx->height ? ist->dec_ctx->height : ist->sub2video.h;
|
|
||||||
|
|
||||||
/* rectangles are AV_PIX_FMT_PAL8, but we have no guarantee that the
|
|
||||||
palettes for all rectangles are identical or compatible */
|
|
||||||
ifp->format = AV_PIX_FMT_RGB32;
|
|
||||||
|
|
||||||
ist->sub2video.frame = av_frame_alloc();
|
ist->sub2video.frame = av_frame_alloc();
|
||||||
if (!ist->sub2video.frame)
|
if (!ist->sub2video.frame)
|
||||||
return AVERROR(ENOMEM);
|
return AVERROR(ENOMEM);
|
||||||
@@ -1525,7 +1493,7 @@ int ifilter_parameters_from_dec(InputFilter *ifilter, const AVCodecContext *dec)
|
|||||||
ifp->fallback.width = dec->width;
|
ifp->fallback.width = dec->width;
|
||||||
ifp->fallback.height = dec->height;
|
ifp->fallback.height = dec->height;
|
||||||
ifp->fallback.sample_aspect_ratio = dec->sample_aspect_ratio;
|
ifp->fallback.sample_aspect_ratio = dec->sample_aspect_ratio;
|
||||||
} else {
|
} else if (dec->codec_type == AVMEDIA_TYPE_AUDIO) {
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
ifp->fallback.format = dec->sample_fmt;
|
ifp->fallback.format = dec->sample_fmt;
|
||||||
@@ -1534,6 +1502,17 @@ int ifilter_parameters_from_dec(InputFilter *ifilter, const AVCodecContext *dec)
|
|||||||
ret = av_channel_layout_copy(&ifp->fallback.ch_layout, &dec->ch_layout);
|
ret = av_channel_layout_copy(&ifp->fallback.ch_layout, &dec->ch_layout);
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
} else {
|
||||||
|
// for subtitles (i.e. sub2video) we set the actual parameters,
|
||||||
|
// rather than just fallback
|
||||||
|
ifp->width = ifp->ist->sub2video.w;
|
||||||
|
ifp->height = ifp->ist->sub2video.h;
|
||||||
|
|
||||||
|
/* rectangles are AV_PIX_FMT_PAL8, but we have no guarantee that the
|
||||||
|
palettes for all rectangles are identical or compatible */
|
||||||
|
ifp->format = AV_PIX_FMT_RGB32;
|
||||||
|
|
||||||
|
av_log(NULL, AV_LOG_VERBOSE, "sub2video: using %dx%d canvas\n", ifp->width, ifp->height);
|
||||||
}
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
@@ -1574,8 +1553,7 @@ int ifilter_has_all_input_formats(FilterGraph *fg)
|
|||||||
int i;
|
int i;
|
||||||
for (i = 0; i < fg->nb_inputs; i++) {
|
for (i = 0; i < fg->nb_inputs; i++) {
|
||||||
InputFilterPriv *ifp = ifp_from_ifilter(fg->inputs[i]);
|
InputFilterPriv *ifp = ifp_from_ifilter(fg->inputs[i]);
|
||||||
if (ifp->format < 0 && (ifp->type_src == AVMEDIA_TYPE_AUDIO ||
|
if (ifp->format < 0)
|
||||||
ifp->type_src == AVMEDIA_TYPE_VIDEO))
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
return 1;
|
return 1;
|
||||||
@@ -1709,9 +1687,7 @@ int ifilter_send_eof(InputFilter *ifilter, int64_t pts, AVRational tb)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (ifp->format < 0 &&
|
if (ifp->format < 0) {
|
||||||
(ifp->type_src == AVMEDIA_TYPE_AUDIO ||
|
|
||||||
ifp->type_src == AVMEDIA_TYPE_VIDEO)) {
|
|
||||||
av_log(NULL, AV_LOG_ERROR,
|
av_log(NULL, AV_LOG_ERROR,
|
||||||
"Cannot determine format of input stream %d:%d after EOF\n",
|
"Cannot determine format of input stream %d:%d after EOF\n",
|
||||||
ifp->ist->file_index, ifp->ist->index);
|
ifp->ist->file_index, ifp->ist->index);
|
||||||
|
Reference in New Issue
Block a user