You've already forked FFmpeg
							
							
				mirror of
				https://github.com/FFmpeg/FFmpeg.git
				synced 2025-10-30 23:18:11 +02:00 
			
		
		
		
	lavf: make overlay_qsv work based on framesync
The existing version which was cherry-picked from Libav does not work
with FFmpeg framework, because ff_request_frame() was totally
different between Libav (recursive) and FFmpeg (non-recursive).
The existing overlay_qsv implementation depends on the recursive version
of ff_request_frame to trigger immediate call to request_frame() on input pad.
But this has been removed in FFmpeg since "lavfi: make request_frame() non-recursive."
Now that we have handy framesync support in FFmpeg, so I make it work
based on framesync. Some other fixing which is also needed to make
overlay_qsv work are put in a separate patch.
Signed-off-by: Ruiling Song <ruiling.song@intel.com>
(cherry picked from commit f3341a0452)
			
			
This commit is contained in:
		
				
					committed by
					
						 Mark Thompson
						Mark Thompson
					
				
			
			
				
	
			
			
			
						parent
						
							6b2fee19a7
						
					
				
				
					commit
					58569162c2
				
			| @@ -267,7 +267,7 @@ OBJS-$(CONFIG_OSCILLOSCOPE_FILTER)           += vf_datascope.o | ||||
| OBJS-$(CONFIG_OVERLAY_FILTER)                += vf_overlay.o framesync.o | ||||
| OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER)         += vf_overlay_opencl.o opencl.o \ | ||||
|                                                 opencl/overlay.o framesync.o | ||||
| OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o | ||||
| OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o framesync.o | ||||
| OBJS-$(CONFIG_OWDENOISE_FILTER)              += vf_owdenoise.o | ||||
| OBJS-$(CONFIG_PAD_FILTER)                    += vf_pad.o | ||||
| OBJS-$(CONFIG_PALETTEGEN_FILTER)             += vf_palettegen.o | ||||
|   | ||||
| @@ -36,6 +36,7 @@ | ||||
| #include "formats.h" | ||||
| #include "video.h" | ||||
|  | ||||
| #include "framesync.h" | ||||
| #include "qsvvpp.h" | ||||
|  | ||||
| #define MAIN    0 | ||||
| @@ -56,14 +57,10 @@ enum var_name { | ||||
|     VAR_VARS_NB | ||||
| }; | ||||
|  | ||||
| enum EOFAction { | ||||
|     EOF_ACTION_REPEAT, | ||||
|     EOF_ACTION_ENDALL | ||||
| }; | ||||
|  | ||||
| typedef struct QSVOverlayContext { | ||||
|     const AVClass      *class; | ||||
|  | ||||
|     FFFrameSync fs; | ||||
|     QSVVPPContext      *qsv; | ||||
|     QSVVPPParam        qsv_param; | ||||
|     mfxExtVPPComposite comp_conf; | ||||
| @@ -72,10 +69,6 @@ typedef struct QSVOverlayContext { | ||||
|     char     *overlay_ox, *overlay_oy, *overlay_ow, *overlay_oh; | ||||
|     uint16_t  overlay_alpha, overlay_pixel_alpha; | ||||
|  | ||||
|     enum EOFAction eof_action;  /* action to take on EOF from source */ | ||||
|  | ||||
|     AVFrame *main; | ||||
|     AVFrame *over_prev, *over_next; | ||||
| } QSVOverlayContext; | ||||
|  | ||||
| static const char *const var_names[] = { | ||||
| @@ -90,20 +83,25 @@ static const char *const var_names[] = { | ||||
|     NULL | ||||
| }; | ||||
|  | ||||
| static const AVOption options[] = { | ||||
| static const AVOption overlay_qsv_options[] = { | ||||
|     { "x", "Overlay x position", OFFSET(overlay_ox), AV_OPT_TYPE_STRING, { .str="0"}, 0, 255, .flags = FLAGS}, | ||||
|     { "y", "Overlay y position", OFFSET(overlay_oy), AV_OPT_TYPE_STRING, { .str="0"}, 0, 255, .flags = FLAGS}, | ||||
|     { "w", "Overlay width",      OFFSET(overlay_ow), AV_OPT_TYPE_STRING, { .str="overlay_iw"}, 0, 255, .flags = FLAGS}, | ||||
|     { "h", "Overlay height",     OFFSET(overlay_oh), AV_OPT_TYPE_STRING, { .str="overlay_ih*w/overlay_iw"}, 0, 255, .flags = FLAGS}, | ||||
|     { "alpha", "Overlay global alpha", OFFSET(overlay_alpha), AV_OPT_TYPE_INT, { .i64 = 255}, 0, 255, .flags = FLAGS}, | ||||
|     { "eof_action", "Action to take when encountering EOF from secondary input ", | ||||
|         OFFSET(eof_action), AV_OPT_TYPE_INT, { .i64 = EOF_ACTION_REPEAT }, | ||||
|         EOF_ACTION_REPEAT, EOF_ACTION_ENDALL, .flags = FLAGS, "eof_action" }, | ||||
|         { "repeat", "Repeat the previous frame.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_REPEAT }, .flags = FLAGS, "eof_action" }, | ||||
|         { "endall", "End both streams.",          0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_ENDALL }, .flags = FLAGS, "eof_action" }, | ||||
|         OFFSET(fs.opt_eof_action), AV_OPT_TYPE_INT, { .i64 = EOF_ACTION_REPEAT }, | ||||
|         EOF_ACTION_REPEAT, EOF_ACTION_PASS, .flags = FLAGS, "eof_action" }, | ||||
|         { "repeat", "Repeat the previous frame.",   0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_REPEAT }, .flags = FLAGS, "eof_action" }, | ||||
|         { "endall", "End both streams.",            0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_ENDALL }, .flags = FLAGS, "eof_action" }, | ||||
|         { "pass",   "Pass through the main input.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_PASS },   .flags = FLAGS, "eof_action" }, | ||||
|     { "shortest", "force termination when the shortest input terminates", OFFSET(fs.opt_shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, | ||||
|     { "repeatlast", "repeat overlay of the last overlay frame", OFFSET(fs.opt_repeatlast), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS }, | ||||
|     { NULL } | ||||
| }; | ||||
|  | ||||
| FRAMESYNC_DEFINE_CLASS(overlay_qsv, QSVOverlayContext, fs); | ||||
|  | ||||
| static int eval_expr(AVFilterContext *ctx) | ||||
| { | ||||
|     QSVOverlayContext *vpp = ctx->priv; | ||||
| @@ -230,12 +228,53 @@ static int config_overlay_input(AVFilterLink *inlink) | ||||
|     return 0; | ||||
| } | ||||
|  | ||||
| static int process_frame(FFFrameSync *fs) | ||||
| { | ||||
|     AVFilterContext  *ctx = fs->parent; | ||||
|     QSVOverlayContext  *s = fs->opaque; | ||||
|     AVFrame        *frame = NULL; | ||||
|     int               ret = 0, i; | ||||
|  | ||||
|     for (i = 0; i < ctx->nb_inputs; i++) { | ||||
|         ret = ff_framesync_get_frame(fs, i, &frame, 0); | ||||
|         if (ret == 0) | ||||
|             ret = ff_qsvvpp_filter_frame(s->qsv, ctx->inputs[i], frame); | ||||
|         if (ret < 0 && ret != AVERROR(EAGAIN)) | ||||
|             break; | ||||
|     } | ||||
|  | ||||
|     return ret; | ||||
| } | ||||
|  | ||||
| static int init_framesync(AVFilterContext *ctx) | ||||
| { | ||||
|     QSVOverlayContext *s = ctx->priv; | ||||
|     int ret, i; | ||||
|  | ||||
|     s->fs.on_event = process_frame; | ||||
|     s->fs.opaque   = s; | ||||
|     ret = ff_framesync_init(&s->fs, ctx, ctx->nb_inputs); | ||||
|     if (ret < 0) | ||||
|         return ret; | ||||
|  | ||||
|     for (i = 0; i < ctx->nb_inputs; i++) { | ||||
|         FFFrameSyncIn *in = &s->fs.in[i]; | ||||
|         in->before    = EXT_STOP; | ||||
|         in->after     = EXT_INFINITY; | ||||
|         in->sync      = i ? 1 : 2; | ||||
|         in->time_base = ctx->inputs[i]->time_base; | ||||
|     } | ||||
|  | ||||
|     return ff_framesync_configure(&s->fs); | ||||
| } | ||||
|  | ||||
| static int config_output(AVFilterLink *outlink) | ||||
| { | ||||
|     AVFilterContext   *ctx = outlink->src; | ||||
|     QSVOverlayContext *vpp = ctx->priv; | ||||
|     AVFilterLink      *in0 = ctx->inputs[0]; | ||||
|     AVFilterLink      *in1 = ctx->inputs[1]; | ||||
|     int ret; | ||||
|  | ||||
|     av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format)); | ||||
|     if ((in0->format == AV_PIX_FMT_QSV && in1->format != AV_PIX_FMT_QSV) || | ||||
| @@ -257,121 +296,27 @@ static int config_output(AVFilterLink *outlink) | ||||
|     outlink->frame_rate = in0->frame_rate; | ||||
|     outlink->time_base  = av_inv_q(outlink->frame_rate); | ||||
|  | ||||
|     ret = init_framesync(ctx); | ||||
|     if (ret < 0) | ||||
|         return ret; | ||||
|  | ||||
|     return ff_qsvvpp_create(ctx, &vpp->qsv, &vpp->qsv_param); | ||||
| } | ||||
|  | ||||
| static int blend_frame(AVFilterContext *ctx, AVFrame *mpic, AVFrame *opic) | ||||
| /* | ||||
|  * Callback for qsvvpp | ||||
|  * @Note: qsvvpp composition does not generate PTS for result frame. | ||||
|  *        so we assign the PTS from framesync to the output frame. | ||||
|  */ | ||||
|  | ||||
| static int filter_callback(AVFilterLink *outlink, AVFrame *frame) | ||||
| { | ||||
|     int                ret = 0; | ||||
|     QSVOverlayContext *vpp = ctx->priv; | ||||
|     AVFrame     *opic_copy = NULL; | ||||
|  | ||||
|     ret = ff_qsvvpp_filter_frame(vpp->qsv, ctx->inputs[0], mpic); | ||||
|     if (ret == 0 || ret == AVERROR(EAGAIN)) { | ||||
|         /* Reference the overlay frame. Because: | ||||
|          * 1. ff_qsvvpp_filter_frame will take control of the given frame | ||||
|          * 2. We need to repeat the overlay frame when 2nd input goes into EOF | ||||
|          */ | ||||
|         opic_copy = av_frame_clone(opic); | ||||
|         if (!opic_copy) | ||||
|             return AVERROR(ENOMEM); | ||||
|  | ||||
|         ret = ff_qsvvpp_filter_frame(vpp->qsv, ctx->inputs[1], opic_copy); | ||||
|     } | ||||
|  | ||||
|     return ret; | ||||
|     QSVOverlayContext *s = outlink->src->priv; | ||||
|     frame->pts = av_rescale_q(s->fs.pts, | ||||
|                               s->fs.time_base, outlink->time_base); | ||||
|     return ff_filter_frame(outlink, frame); | ||||
| } | ||||
|  | ||||
| static int handle_overlay_eof(AVFilterContext *ctx) | ||||
| { | ||||
|     int              ret = 0; | ||||
|     QSVOverlayContext *s = ctx->priv; | ||||
|     /* Repeat previous frame on secondary input */ | ||||
|     if (s->over_prev && s->eof_action == EOF_ACTION_REPEAT) | ||||
|         ret = blend_frame(ctx, s->main, s->over_prev); | ||||
|     /* End both streams */ | ||||
|     else if (s->eof_action == EOF_ACTION_ENDALL) | ||||
|         return AVERROR_EOF; | ||||
|  | ||||
|     s->main = NULL; | ||||
|  | ||||
|     return ret; | ||||
| } | ||||
|  | ||||
| static int request_frame(AVFilterLink *outlink) | ||||
| { | ||||
|     AVFilterContext *ctx = outlink->src; | ||||
|     QSVOverlayContext *s = ctx->priv; | ||||
|     AVRational   tb_main = ctx->inputs[MAIN]->time_base; | ||||
|     AVRational   tb_over = ctx->inputs[OVERLAY]->time_base; | ||||
|     int              ret = 0; | ||||
|  | ||||
|     /* get a frame on the main input */ | ||||
|     if (!s->main) { | ||||
|         ret = ff_request_frame(ctx->inputs[MAIN]); | ||||
|         if (ret < 0) | ||||
|             return ret; | ||||
|     } | ||||
|  | ||||
|     /* get a new frame on the overlay input, on EOF check setting 'eof_action' */ | ||||
|     if (!s->over_next) { | ||||
|         ret = ff_request_frame(ctx->inputs[OVERLAY]); | ||||
|         if (ret == AVERROR_EOF) | ||||
|             return handle_overlay_eof(ctx); | ||||
|         else if (ret < 0) | ||||
|             return ret; | ||||
|     } | ||||
|  | ||||
|     while (s->main->pts != AV_NOPTS_VALUE && | ||||
|            s->over_next->pts != AV_NOPTS_VALUE && | ||||
|            av_compare_ts(s->over_next->pts, tb_over, s->main->pts, tb_main) < 0) { | ||||
|         av_frame_free(&s->over_prev); | ||||
|         FFSWAP(AVFrame*, s->over_prev, s->over_next); | ||||
|  | ||||
|         ret = ff_request_frame(ctx->inputs[OVERLAY]); | ||||
|         if (ret == AVERROR_EOF) | ||||
|             return handle_overlay_eof(ctx); | ||||
|         else if (ret < 0) | ||||
|             return ret; | ||||
|     } | ||||
|  | ||||
|     if (s->main->pts == AV_NOPTS_VALUE || | ||||
|         s->over_next->pts == AV_NOPTS_VALUE || | ||||
|         !av_compare_ts(s->over_next->pts, tb_over, s->main->pts, tb_main)) { | ||||
|         ret = blend_frame(ctx, s->main, s->over_next); | ||||
|         av_frame_free(&s->over_prev); | ||||
|         FFSWAP(AVFrame*, s->over_prev, s->over_next); | ||||
|     } else if (s->over_prev) { | ||||
|         ret = blend_frame(ctx, s->main, s->over_prev); | ||||
|     } else { | ||||
|         av_frame_free(&s->main); | ||||
|         ret = AVERROR(EAGAIN); | ||||
|     } | ||||
|  | ||||
|     s->main = NULL; | ||||
|  | ||||
|     return ret; | ||||
| } | ||||
|  | ||||
| static int filter_frame_main(AVFilterLink *inlink, AVFrame *frame) | ||||
| { | ||||
|     QSVOverlayContext *s = inlink->dst->priv; | ||||
|  | ||||
|     av_assert0(!s->main); | ||||
|     s->main = frame; | ||||
|  | ||||
|     return 0; | ||||
| } | ||||
|  | ||||
| static int filter_frame_overlay(AVFilterLink *inlink, AVFrame *frame) | ||||
| { | ||||
|     QSVOverlayContext *s = inlink->dst->priv; | ||||
|  | ||||
|     av_assert0(!s->over_next); | ||||
|     s->over_next = frame; | ||||
|  | ||||
|     return 0; | ||||
| } | ||||
|  | ||||
| static int overlay_qsv_init(AVFilterContext *ctx) | ||||
| { | ||||
| @@ -387,7 +332,7 @@ static int overlay_qsv_init(AVFilterContext *ctx) | ||||
|         return AVERROR(ENOMEM); | ||||
|  | ||||
|     /* initialize QSVVPP params */ | ||||
|     vpp->qsv_param.filter_frame = NULL; | ||||
|     vpp->qsv_param.filter_frame = filter_callback; | ||||
|     vpp->qsv_param.ext_buf      = av_mallocz(sizeof(*vpp->qsv_param.ext_buf)); | ||||
|     if (!vpp->qsv_param.ext_buf) | ||||
|         return AVERROR(ENOMEM); | ||||
| @@ -404,14 +349,18 @@ static void overlay_qsv_uninit(AVFilterContext *ctx) | ||||
| { | ||||
|     QSVOverlayContext *vpp = ctx->priv; | ||||
|  | ||||
|     av_frame_free(&vpp->main); | ||||
|     av_frame_free(&vpp->over_prev); | ||||
|     av_frame_free(&vpp->over_next); | ||||
|     ff_qsvvpp_free(&vpp->qsv); | ||||
|     ff_framesync_uninit(&vpp->fs); | ||||
|     av_freep(&vpp->comp_conf.InputStream); | ||||
|     av_freep(&vpp->qsv_param.ext_buf); | ||||
| } | ||||
|  | ||||
| static int activate(AVFilterContext *ctx) | ||||
| { | ||||
|     QSVOverlayContext *s = ctx->priv; | ||||
|     return ff_framesync_activate(&s->fs); | ||||
| } | ||||
|  | ||||
| static int overlay_qsv_query_formats(AVFilterContext *ctx) | ||||
| { | ||||
|     int i; | ||||
| @@ -444,25 +393,16 @@ static int overlay_qsv_query_formats(AVFilterContext *ctx) | ||||
|     return 0; | ||||
| } | ||||
|  | ||||
| static const AVClass overlay_qsv_class = { | ||||
|     .class_name = "overlay_qsv", | ||||
|     .item_name  = av_default_item_name, | ||||
|     .option     = options, | ||||
|     .version    = LIBAVUTIL_VERSION_INT, | ||||
| }; | ||||
|  | ||||
| static const AVFilterPad overlay_qsv_inputs[] = { | ||||
|     { | ||||
|         .name          = "main", | ||||
|         .type          = AVMEDIA_TYPE_VIDEO, | ||||
|         .filter_frame  = filter_frame_main, | ||||
|         .config_props  = config_main_input, | ||||
|         .needs_fifo    = 1, | ||||
|     }, | ||||
|     { | ||||
|         .name          = "overlay", | ||||
|         .type          = AVMEDIA_TYPE_VIDEO, | ||||
|         .filter_frame  = filter_frame_overlay, | ||||
|         .config_props  = config_overlay_input, | ||||
|         .needs_fifo    = 1, | ||||
|     }, | ||||
| @@ -474,7 +414,6 @@ static const AVFilterPad overlay_qsv_outputs[] = { | ||||
|         .name          = "default", | ||||
|         .type          = AVMEDIA_TYPE_VIDEO, | ||||
|         .config_props  = config_output, | ||||
|         .request_frame = request_frame, | ||||
|     }, | ||||
|     { NULL } | ||||
| }; | ||||
| @@ -484,8 +423,10 @@ AVFilter ff_vf_overlay_qsv = { | ||||
|     .description    = NULL_IF_CONFIG_SMALL("Quick Sync Video overlay."), | ||||
|     .priv_size      = sizeof(QSVOverlayContext), | ||||
|     .query_formats  = overlay_qsv_query_formats, | ||||
|     .preinit        = overlay_qsv_framesync_preinit, | ||||
|     .init           = overlay_qsv_init, | ||||
|     .uninit         = overlay_qsv_uninit, | ||||
|     .activate       = activate, | ||||
|     .inputs         = overlay_qsv_inputs, | ||||
|     .outputs        = overlay_qsv_outputs, | ||||
|     .priv_class     = &overlay_qsv_class, | ||||
|   | ||||
		Reference in New Issue
	
	Block a user