1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00
FFmpeg/libavfilter/vf_hwupload.c
Mark Thompson c6bc18bc12 vf_hwupload/hwmap: Support setting a fixed pool size
These filters do not directly know whether the API they are using will
support dynamic frame pools, so this is somewhat tricky.  If the user
sets extra_hw_frames, we assume that they are aware of the problem and
set a fixed size based on that.  If not, most cases use dynamic sizing
just like they did previously.  The hardware-reverse-mapping case for
hwmap previously had a large fixed size (64) here, primarily as a hack
for QSV use - this is removed and extra_hw_frames will need to be set
for QSV to work since it requires fixed-size pools (as the other cases
do, and which didn't work before).
2018-02-11 22:11:10 +00:00

241 lines
6.7 KiB
C

/*
* This file is part of Libav.
*
* Libav is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* Libav is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with Libav; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/buffer.h"
#include "libavutil/hwcontext.h"
#include "libavutil/hwcontext_internal.h"
#include "libavutil/log.h"
#include "libavutil/pixdesc.h"
#include "libavutil/opt.h"
#include "avfilter.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
typedef struct HWUploadContext {
const AVClass *class;
AVBufferRef *hwdevice_ref;
AVHWDeviceContext *hwdevice;
AVBufferRef *hwframes_ref;
AVHWFramesContext *hwframes;
} HWUploadContext;
static int hwupload_query_formats(AVFilterContext *avctx)
{
HWUploadContext *ctx = avctx->priv;
AVHWFramesConstraints *constraints = NULL;
const enum AVPixelFormat *input_pix_fmts, *output_pix_fmts;
AVFilterFormats *input_formats = NULL;
int err, i;
if (!avctx->hw_device_ctx) {
av_log(ctx, AV_LOG_ERROR, "A hardware device reference is required "
"to upload frames to.\n");
return AVERROR(EINVAL);
}
ctx->hwdevice_ref = av_buffer_ref(avctx->hw_device_ctx);
if (!ctx->hwdevice_ref)
return AVERROR(ENOMEM);
ctx->hwdevice = (AVHWDeviceContext*)ctx->hwdevice_ref->data;
constraints = av_hwdevice_get_hwframe_constraints(ctx->hwdevice_ref, NULL);
if (!constraints) {
err = AVERROR(EINVAL);
goto fail;
}
input_pix_fmts = constraints->valid_sw_formats;
output_pix_fmts = constraints->valid_hw_formats;
input_formats = ff_make_format_list(output_pix_fmts);
if (!input_formats) {
err = AVERROR(ENOMEM);
goto fail;
}
if (input_pix_fmts) {
for (i = 0; input_pix_fmts[i] != AV_PIX_FMT_NONE; i++) {
err = ff_add_format(&input_formats, input_pix_fmts[i]);
if (err < 0) {
ff_formats_unref(&input_formats);
goto fail;
}
}
}
ff_formats_ref(input_formats, &avctx->inputs[0]->out_formats);
ff_formats_ref(ff_make_format_list(output_pix_fmts),
&avctx->outputs[0]->in_formats);
av_hwframe_constraints_free(&constraints);
return 0;
fail:
av_buffer_unref(&ctx->hwdevice_ref);
av_hwframe_constraints_free(&constraints);
return err;
}
static int hwupload_config_output(AVFilterLink *outlink)
{
AVFilterContext *avctx = outlink->src;
AVFilterLink *inlink = avctx->inputs[0];
HWUploadContext *ctx = avctx->priv;
int err;
av_buffer_unref(&ctx->hwframes_ref);
if (inlink->format == outlink->format) {
// The input is already a hardware format, so we just want to
// pass through the input frames in their own hardware context.
if (!inlink->hw_frames_ctx) {
av_log(ctx, AV_LOG_ERROR, "No input hwframe context.\n");
return AVERROR(EINVAL);
}
outlink->hw_frames_ctx = av_buffer_ref(inlink->hw_frames_ctx);
if (!outlink->hw_frames_ctx)
return AVERROR(ENOMEM);
return 0;
}
ctx->hwframes_ref = av_hwframe_ctx_alloc(ctx->hwdevice_ref);
if (!ctx->hwframes_ref)
return AVERROR(ENOMEM);
ctx->hwframes = (AVHWFramesContext*)ctx->hwframes_ref->data;
av_log(ctx, AV_LOG_DEBUG, "Surface format is %s.\n",
av_get_pix_fmt_name(inlink->format));
ctx->hwframes->format = outlink->format;
ctx->hwframes->sw_format = inlink->format;
ctx->hwframes->width = inlink->w;
ctx->hwframes->height = inlink->h;
if (avctx->extra_hw_frames >= 0)
ctx->hwframes->initial_pool_size = 2 + avctx->extra_hw_frames;
err = av_hwframe_ctx_init(ctx->hwframes_ref);
if (err < 0)
goto fail;
outlink->hw_frames_ctx = av_buffer_ref(ctx->hwframes_ref);
if (!outlink->hw_frames_ctx) {
err = AVERROR(ENOMEM);
goto fail;
}
return 0;
fail:
av_buffer_unref(&ctx->hwframes_ref);
return err;
}
static int hwupload_filter_frame(AVFilterLink *link, AVFrame *input)
{
AVFilterContext *avctx = link->dst;
AVFilterLink *outlink = avctx->outputs[0];
HWUploadContext *ctx = avctx->priv;
AVFrame *output = NULL;
int err;
if (input->format == outlink->format)
return ff_filter_frame(outlink, input);
output = ff_get_video_buffer(outlink, outlink->w, outlink->h);
if (!output) {
av_log(ctx, AV_LOG_ERROR, "Failed to allocate frame to upload to.\n");
err = AVERROR(ENOMEM);
goto fail;
}
output->width = input->width;
output->height = input->height;
err = av_hwframe_transfer_data(output, input, 0);
if (err < 0) {
av_log(ctx, AV_LOG_ERROR, "Failed to upload frame: %d.\n", err);
goto fail;
}
err = av_frame_copy_props(output, input);
if (err < 0)
goto fail;
av_frame_free(&input);
return ff_filter_frame(outlink, output);
fail:
av_frame_free(&input);
av_frame_free(&output);
return err;
}
static av_cold void hwupload_uninit(AVFilterContext *avctx)
{
HWUploadContext *ctx = avctx->priv;
av_buffer_unref(&ctx->hwframes_ref);
av_buffer_unref(&ctx->hwdevice_ref);
}
static const AVClass hwupload_class = {
.class_name = "hwupload",
.item_name = av_default_item_name,
.option = NULL,
.version = LIBAVUTIL_VERSION_INT,
};
static const AVFilterPad hwupload_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = hwupload_filter_frame,
},
{ NULL }
};
static const AVFilterPad hwupload_outputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = hwupload_config_output,
},
{ NULL }
};
AVFilter ff_vf_hwupload = {
.name = "hwupload",
.description = NULL_IF_CONFIG_SMALL("Upload a normal frame to a hardware frame"),
.uninit = hwupload_uninit,
.query_formats = hwupload_query_formats,
.priv_size = sizeof(HWUploadContext),
.priv_class = &hwupload_class,
.inputs = hwupload_inputs,
.outputs = hwupload_outputs,
.flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE,
};