1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-21 10:55:51 +02:00
FFmpeg/libavcodec/d3d12va_decode.c
Tong Wu 0faf2ca98a avcodec/d3d12va_decode: remove extra spaces for declaration
Signed-off-by: Tong Wu <tong1.wu@intel.com>
2024-04-03 10:23:05 +08:00

538 lines
18 KiB
C

/*
* Direct3D 12 HW acceleration video decoder
*
* copyright (c) 2022-2023 Wu Jianhua <toqsxw@outlook.com>
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <string.h>
#include <initguid.h>
#include "libavutil/common.h"
#include "libavutil/log.h"
#include "libavutil/mem.h"
#include "libavutil/time.h"
#include "libavutil/imgutils.h"
#include "libavutil/hwcontext_d3d12va_internal.h"
#include "libavutil/hwcontext_d3d12va.h"
#include "avcodec.h"
#include "decode.h"
#include "d3d12va_decode.h"
#include "dxva2_internal.h"
typedef struct HelperObjects {
ID3D12CommandAllocator *command_allocator;
ID3D12Resource *buffer;
uint64_t fence_value;
} HelperObjects;
int ff_d3d12va_get_suitable_max_bitstream_size(AVCodecContext *avctx)
{
AVHWFramesContext *frames_ctx = D3D12VA_FRAMES_CONTEXT(avctx);
return av_image_get_buffer_size(frames_ctx->sw_format, avctx->coded_width, avctx->coded_height, 1);
}
unsigned ff_d3d12va_get_surface_index(const AVCodecContext *avctx,
D3D12VADecodeContext *ctx, const AVFrame *frame,
int curr)
{
AVD3D12VAFrame *f;
ID3D12Resource *res;
unsigned i;
f = (AVD3D12VAFrame *)frame->data[0];
if (!f)
goto fail;
res = f->texture;
if (!res)
goto fail;
for (i = 0; i < ctx->max_num_ref; i++) {
if (ctx->ref_resources[i] && res == ctx->ref_resources[i]) {
ctx->used_mask |= 1 << i;
return i;
}
}
if (curr) {
for (i = 0; i < ctx->max_num_ref; i++) {
if (!((ctx->used_mask >> i) & 0x1)) {
ctx->ref_resources[i] = res;
return i;
}
}
}
fail:
av_log((AVCodecContext *)avctx, AV_LOG_WARNING, "Could not get surface index. Using 0 instead.\n");
return 0;
}
static int d3d12va_get_valid_helper_objects(AVCodecContext *avctx, ID3D12CommandAllocator **ppAllocator,
ID3D12Resource **ppBuffer)
{
HRESULT hr;
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
HelperObjects obj = { 0 };
D3D12_HEAP_PROPERTIES heap_props = { .Type = D3D12_HEAP_TYPE_UPLOAD };
D3D12_RESOURCE_DESC desc = {
.Dimension = D3D12_RESOURCE_DIMENSION_BUFFER,
.Alignment = D3D12_DEFAULT_RESOURCE_PLACEMENT_ALIGNMENT,
.Width = ctx->bitstream_size,
.Height = 1,
.DepthOrArraySize = 1,
.MipLevels = 1,
.Format = DXGI_FORMAT_UNKNOWN,
.SampleDesc = { .Count = 1, .Quality = 0 },
.Layout = D3D12_TEXTURE_LAYOUT_ROW_MAJOR,
.Flags = D3D12_RESOURCE_FLAG_NONE,
};
if (av_fifo_peek(ctx->objects_queue, &obj, 1, 0) >= 0) {
uint64_t completion = ID3D12Fence_GetCompletedValue(ctx->sync_ctx.fence);
if (completion >= obj.fence_value) {
*ppAllocator = obj.command_allocator;
*ppBuffer = obj.buffer;
av_fifo_read(ctx->objects_queue, &obj, 1);
return 0;
}
}
hr = ID3D12Device_CreateCommandAllocator(ctx->device_ctx->device, D3D12_COMMAND_LIST_TYPE_VIDEO_DECODE,
&IID_ID3D12CommandAllocator, (void **)ppAllocator);
if (FAILED(hr)) {
av_log(avctx, AV_LOG_ERROR, "Failed to create a new command allocator!\n");
return AVERROR(EINVAL);
}
hr = ID3D12Device_CreateCommittedResource(ctx->device_ctx->device, &heap_props, D3D12_HEAP_FLAG_NONE,
&desc, D3D12_RESOURCE_STATE_GENERIC_READ, NULL,
&IID_ID3D12Resource, (void **)ppBuffer);
if (FAILED(hr)) {
av_log(avctx, AV_LOG_ERROR, "Failed to create a new d3d12 buffer!\n");
return AVERROR(EINVAL);
}
return 0;
}
static int d3d12va_discard_helper_objects(AVCodecContext *avctx, ID3D12CommandAllocator *pAllocator,
ID3D12Resource *pBuffer, uint64_t fence_value)
{
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
HelperObjects obj = {
.command_allocator = pAllocator,
.buffer = pBuffer,
.fence_value = fence_value,
};
if (av_fifo_write(ctx->objects_queue, &obj, 1) < 0) {
D3D12_OBJECT_RELEASE(pAllocator);
D3D12_OBJECT_RELEASE(pBuffer);
return AVERROR(ENOMEM);
}
return 0;
}
static int d3d12va_fence_completion(AVD3D12VASyncContext *psync_ctx)
{
uint64_t completion = ID3D12Fence_GetCompletedValue(psync_ctx->fence);
if (completion < psync_ctx->fence_value) {
if (FAILED(ID3D12Fence_SetEventOnCompletion(psync_ctx->fence, psync_ctx->fence_value, psync_ctx->event)))
return AVERROR(EINVAL);
WaitForSingleObjectEx(psync_ctx->event, INFINITE, FALSE);
}
return 0;
}
static void bufref_free_interface(void *opaque, uint8_t *data)
{
D3D12_OBJECT_RELEASE(opaque);
}
static AVBufferRef *bufref_wrap_interface(IUnknown *iface)
{
return av_buffer_create((uint8_t*)iface, 1, bufref_free_interface, iface, 0);
}
static int d3d12va_sync_with_gpu(AVCodecContext *avctx)
{
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
DX_CHECK(ID3D12CommandQueue_Signal(ctx->command_queue, ctx->sync_ctx.fence, ++ctx->sync_ctx.fence_value));
return d3d12va_fence_completion(&ctx->sync_ctx);
fail:
return AVERROR(EINVAL);
}
static int d3d12va_create_decoder_heap(AVCodecContext *avctx)
{
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
AVHWFramesContext *frames_ctx = D3D12VA_FRAMES_CONTEXT(avctx);
AVD3D12VADeviceContext *device_hwctx = ctx->device_ctx;
AVD3D12VAFramesContext *frames_hwctx = frames_ctx->hwctx;
D3D12_VIDEO_DECODER_HEAP_DESC desc = {
.NodeMask = 0,
.Configuration = ctx->cfg,
.DecodeWidth = frames_ctx->width,
.DecodeHeight = frames_ctx->height,
.Format = frames_hwctx->format,
.FrameRate = { avctx->framerate.num, avctx->framerate.den },
.BitRate = avctx->bit_rate,
.MaxDecodePictureBufferCount = ctx->max_num_ref,
};
DX_CHECK(ID3D12VideoDevice_CreateVideoDecoderHeap(device_hwctx->video_device, &desc,
&IID_ID3D12VideoDecoderHeap, (void **)&ctx->decoder_heap));
return 0;
fail:
if (ctx->decoder) {
av_log(avctx, AV_LOG_ERROR, "D3D12 doesn't support decoding frames with an extent "
"[width(%d), height(%d)], on your device!\n", frames_ctx->width, frames_ctx->height);
}
return AVERROR(EINVAL);
}
static int d3d12va_create_decoder(AVCodecContext *avctx)
{
D3D12_VIDEO_DECODER_DESC desc;
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
AVHWFramesContext *frames_ctx = D3D12VA_FRAMES_CONTEXT(avctx);
AVD3D12VADeviceContext *device_hwctx = ctx->device_ctx;
AVD3D12VAFramesContext *frames_hwctx = frames_ctx->hwctx;
D3D12_FEATURE_DATA_VIDEO_DECODE_SUPPORT feature = {
.NodeIndex = 0,
.Configuration = ctx->cfg,
.Width = frames_ctx->width,
.Height = frames_ctx->height,
.DecodeFormat = frames_hwctx->format,
.FrameRate = { avctx->framerate.num, avctx->framerate.den },
.BitRate = avctx->bit_rate,
};
DX_CHECK(ID3D12VideoDevice_CheckFeatureSupport(device_hwctx->video_device, D3D12_FEATURE_VIDEO_DECODE_SUPPORT,
&feature, sizeof(feature)));
if (!(feature.SupportFlags & D3D12_VIDEO_DECODE_SUPPORT_FLAG_SUPPORTED)) {
av_log(avctx, AV_LOG_ERROR, "D3D12 video decode is not supported on this device.\n");
return AVERROR(ENOSYS);
}
if (!(feature.DecodeTier >= D3D12_VIDEO_DECODE_TIER_2)) {
av_log(avctx, AV_LOG_ERROR, "D3D12 video decode on this device requires tier %d support, "
"but it is not implemented.\n", feature.DecodeTier);
return AVERROR_PATCHWELCOME;
}
desc = (D3D12_VIDEO_DECODER_DESC) {
.NodeMask = 0,
.Configuration = ctx->cfg,
};
DX_CHECK(ID3D12VideoDevice_CreateVideoDecoder(device_hwctx->video_device, &desc, &IID_ID3D12VideoDecoder,
(void **)&ctx->decoder));
ctx->decoder_ref = bufref_wrap_interface((IUnknown *)ctx->decoder);
if (!ctx->decoder_ref)
return AVERROR(ENOMEM);
return 0;
fail:
return AVERROR(EINVAL);
}
int ff_d3d12va_common_frame_params(AVCodecContext *avctx, AVBufferRef *hw_frames_ctx)
{
AVHWFramesContext *frames_ctx = (AVHWFramesContext *)hw_frames_ctx->data;
frames_ctx->format = AV_PIX_FMT_D3D12;
frames_ctx->sw_format = avctx->sw_pix_fmt == AV_PIX_FMT_YUV420P10 ? AV_PIX_FMT_P010 : AV_PIX_FMT_NV12;
frames_ctx->width = avctx->width;
frames_ctx->height = avctx->height;
return 0;
}
int ff_d3d12va_decode_init(AVCodecContext *avctx)
{
int ret;
AVHWFramesContext *frames_ctx;
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
ID3D12Resource *buffer = NULL;
ID3D12CommandAllocator *command_allocator = NULL;
D3D12_COMMAND_QUEUE_DESC queue_desc = {
.Type = D3D12_COMMAND_LIST_TYPE_VIDEO_DECODE,
.Priority = 0,
.Flags = D3D12_COMMAND_QUEUE_FLAG_NONE,
.NodeMask = 0,
};
ctx->pix_fmt = avctx->hwaccel->pix_fmt;
ret = ff_decode_get_hw_frames_ctx(avctx, AV_HWDEVICE_TYPE_D3D12VA);
if (ret < 0)
return ret;
frames_ctx = D3D12VA_FRAMES_CONTEXT(avctx);
ctx->device_ctx = (AVD3D12VADeviceContext *)frames_ctx->device_ctx->hwctx;
if (frames_ctx->format != ctx->pix_fmt) {
av_log(avctx, AV_LOG_ERROR, "Invalid pixfmt for hwaccel!\n");
goto fail;
}
ret = d3d12va_create_decoder(avctx);
if (ret < 0)
goto fail;
ret = d3d12va_create_decoder_heap(avctx);
if (ret < 0)
goto fail;
ctx->bitstream_size = ff_d3d12va_get_suitable_max_bitstream_size(avctx);
ctx->ref_resources = av_calloc(ctx->max_num_ref, sizeof(*ctx->ref_resources));
if (!ctx->ref_resources)
return AVERROR(ENOMEM);
ctx->ref_subresources = av_calloc(ctx->max_num_ref, sizeof(*ctx->ref_subresources));
if (!ctx->ref_subresources)
return AVERROR(ENOMEM);
ctx->objects_queue = av_fifo_alloc2(D3D12VA_VIDEO_DEC_ASYNC_DEPTH,
sizeof(HelperObjects), AV_FIFO_FLAG_AUTO_GROW);
if (!ctx->objects_queue)
return AVERROR(ENOMEM);
DX_CHECK(ID3D12Device_CreateFence(ctx->device_ctx->device, 0, D3D12_FENCE_FLAG_NONE,
&IID_ID3D12Fence, (void **)&ctx->sync_ctx.fence));
ctx->sync_ctx.event = CreateEvent(NULL, FALSE, FALSE, NULL);
if (!ctx->sync_ctx.event)
goto fail;
ret = d3d12va_get_valid_helper_objects(avctx, &command_allocator, &buffer);
if (ret < 0)
goto fail;
DX_CHECK(ID3D12Device_CreateCommandQueue(ctx->device_ctx->device, &queue_desc,
&IID_ID3D12CommandQueue, (void **)&ctx->command_queue));
DX_CHECK(ID3D12Device_CreateCommandList(ctx->device_ctx->device, 0, queue_desc.Type,
command_allocator, NULL, &IID_ID3D12CommandList, (void **)&ctx->command_list));
DX_CHECK(ID3D12VideoDecodeCommandList_Close(ctx->command_list));
ID3D12CommandQueue_ExecuteCommandLists(ctx->command_queue, 1, (ID3D12CommandList **)&ctx->command_list);
ret = d3d12va_sync_with_gpu(avctx);
if (ret < 0)
goto fail;
d3d12va_discard_helper_objects(avctx, command_allocator, buffer, ctx->sync_ctx.fence_value);
if (ret < 0)
goto fail;
return 0;
fail:
D3D12_OBJECT_RELEASE(command_allocator);
D3D12_OBJECT_RELEASE(buffer);
ff_d3d12va_decode_uninit(avctx);
return AVERROR(EINVAL);
}
int ff_d3d12va_decode_uninit(AVCodecContext *avctx)
{
int num_allocator = 0;
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
HelperObjects obj;
if (ctx->sync_ctx.fence)
d3d12va_sync_with_gpu(avctx);
av_freep(&ctx->ref_resources);
av_freep(&ctx->ref_subresources);
D3D12_OBJECT_RELEASE(ctx->command_list);
D3D12_OBJECT_RELEASE(ctx->command_queue);
if (ctx->objects_queue) {
while (av_fifo_read(ctx->objects_queue, &obj, 1) >= 0) {
num_allocator++;
D3D12_OBJECT_RELEASE(obj.buffer);
D3D12_OBJECT_RELEASE(obj.command_allocator);
}
av_log(avctx, AV_LOG_VERBOSE, "Total number of command allocators reused: %d\n", num_allocator);
}
av_fifo_freep2(&ctx->objects_queue);
D3D12_OBJECT_RELEASE(ctx->sync_ctx.fence);
if (ctx->sync_ctx.event)
CloseHandle(ctx->sync_ctx.event);
D3D12_OBJECT_RELEASE(ctx->decoder_heap);
av_buffer_unref(&ctx->decoder_ref);
return 0;
}
static inline int d3d12va_update_reference_frames_state(AVCodecContext *avctx, D3D12_RESOURCE_BARRIER *barriers,
ID3D12Resource *current_resource, int state_before, int state_end)
{
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
int num_barrier = 0;
for (int i = 0; i < ctx->max_num_ref; i++) {
if (((ctx->used_mask >> i) & 0x1) && ctx->ref_resources[i] && ctx->ref_resources[i] != current_resource) {
barriers[num_barrier].Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION;
barriers[num_barrier].Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE;
barriers[num_barrier].Transition = (D3D12_RESOURCE_TRANSITION_BARRIER){
.pResource = ctx->ref_resources[i],
.Subresource = D3D12_RESOURCE_BARRIER_ALL_SUBRESOURCES,
.StateBefore = state_before,
.StateAfter = state_end,
};
num_barrier++;
}
}
return num_barrier;
}
int ff_d3d12va_common_end_frame(AVCodecContext *avctx, AVFrame *frame,
const void *pp, unsigned pp_size,
const void *qm, unsigned qm_size,
int(*update_input_arguments)(AVCodecContext *, D3D12_VIDEO_DECODE_INPUT_STREAM_ARGUMENTS *, ID3D12Resource *))
{
int ret;
D3D12VADecodeContext *ctx = D3D12VA_DECODE_CONTEXT(avctx);
ID3D12Resource *buffer = NULL;
ID3D12CommandAllocator *command_allocator = NULL;
AVD3D12VAFrame *f = (AVD3D12VAFrame *)frame->data[0];
ID3D12Resource *resource = (ID3D12Resource *)f->texture;
ID3D12VideoDecodeCommandList *cmd_list = ctx->command_list;
D3D12_RESOURCE_BARRIER barriers[32] = { 0 };
D3D12_VIDEO_DECODE_INPUT_STREAM_ARGUMENTS input_args = {
.NumFrameArguments = 2,
.FrameArguments = {
[0] = {
.Type = D3D12_VIDEO_DECODE_ARGUMENT_TYPE_PICTURE_PARAMETERS,
.Size = pp_size,
.pData = (void *)pp,
},
[1] = {
.Type = D3D12_VIDEO_DECODE_ARGUMENT_TYPE_INVERSE_QUANTIZATION_MATRIX,
.Size = qm_size,
.pData = (void *)qm,
},
},
.pHeap = ctx->decoder_heap,
};
D3D12_VIDEO_DECODE_OUTPUT_STREAM_ARGUMENTS output_args = {
.ConversionArguments = { 0 },
.OutputSubresource = 0,
.pOutputTexture2D = resource,
};
UINT num_barrier = 1;
barriers[0] = (D3D12_RESOURCE_BARRIER) {
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION,
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE,
.Transition = {
.pResource = resource,
.Subresource = D3D12_RESOURCE_BARRIER_ALL_SUBRESOURCES,
.StateBefore = D3D12_RESOURCE_STATE_COMMON,
.StateAfter = D3D12_RESOURCE_STATE_VIDEO_DECODE_WRITE,
},
};
memset(ctx->ref_subresources, 0, sizeof(UINT) * ctx->max_num_ref);
input_args.ReferenceFrames.NumTexture2Ds = ctx->max_num_ref;
input_args.ReferenceFrames.ppTexture2Ds = ctx->ref_resources;
input_args.ReferenceFrames.pSubresources = ctx->ref_subresources;
ret = d3d12va_fence_completion(&f->sync_ctx);
if (ret < 0)
goto fail;
if (!qm)
input_args.NumFrameArguments = 1;
ret = d3d12va_get_valid_helper_objects(avctx, &command_allocator, &buffer);
if (ret < 0)
goto fail;
ret = update_input_arguments(avctx, &input_args, buffer);
if (ret < 0)
goto fail;
DX_CHECK(ID3D12CommandAllocator_Reset(command_allocator));
DX_CHECK(ID3D12VideoDecodeCommandList_Reset(cmd_list, command_allocator));
num_barrier += d3d12va_update_reference_frames_state(avctx, &barriers[1], resource, D3D12_RESOURCE_STATE_COMMON, D3D12_RESOURCE_STATE_VIDEO_DECODE_READ);
ID3D12VideoDecodeCommandList_ResourceBarrier(cmd_list, num_barrier, barriers);
ID3D12VideoDecodeCommandList_DecodeFrame(cmd_list, ctx->decoder, &output_args, &input_args);
for (int i = 0; i < num_barrier; i++)
FFSWAP(D3D12_RESOURCE_STATES, barriers[i].Transition.StateBefore, barriers[i].Transition.StateAfter);
ID3D12VideoDecodeCommandList_ResourceBarrier(cmd_list, num_barrier, barriers);
DX_CHECK(ID3D12VideoDecodeCommandList_Close(cmd_list));
ID3D12CommandQueue_ExecuteCommandLists(ctx->command_queue, 1, (ID3D12CommandList **)&ctx->command_list);
DX_CHECK(ID3D12CommandQueue_Signal(ctx->command_queue, f->sync_ctx.fence, ++f->sync_ctx.fence_value));
DX_CHECK(ID3D12CommandQueue_Signal(ctx->command_queue, ctx->sync_ctx.fence, ++ctx->sync_ctx.fence_value));
ret = d3d12va_discard_helper_objects(avctx, command_allocator, buffer, ctx->sync_ctx.fence_value);
if (ret < 0)
return ret;
return 0;
fail:
if (command_allocator)
d3d12va_discard_helper_objects(avctx, command_allocator, buffer, ctx->sync_ctx.fence_value);
return AVERROR(EINVAL);
}