mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-18 03:19:31 +02:00
5fc17edc7d
* commit '1520c6ff05d835da4b793318fc88bbbc129c86a1': nvenc: export CPB props side data Merged-by: Hendrik Leppkes <h.leppkes@gmail.com>
1575 lines
53 KiB
C
1575 lines
53 KiB
C
/*
|
|
* H.264 hardware encoding using nvidia nvenc
|
|
* Copyright (c) 2014 Timo Rothenpieler <timo@rothenpieler.org>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#if defined(_WIN32)
|
|
#include <windows.h>
|
|
#else
|
|
#include <dlfcn.h>
|
|
#endif
|
|
|
|
#include <nvEncodeAPI.h>
|
|
|
|
#include "libavutil/internal.h"
|
|
#include "libavutil/imgutils.h"
|
|
#include "libavutil/avassert.h"
|
|
#include "libavutil/opt.h"
|
|
#include "libavutil/mem.h"
|
|
#include "avcodec.h"
|
|
#include "internal.h"
|
|
#include "thread.h"
|
|
|
|
#if defined(_WIN32)
|
|
#define CUDAAPI __stdcall
|
|
#else
|
|
#define CUDAAPI
|
|
#endif
|
|
|
|
#if defined(_WIN32)
|
|
#define LOAD_FUNC(l, s) GetProcAddress(l, s)
|
|
#define DL_CLOSE_FUNC(l) FreeLibrary(l)
|
|
#else
|
|
#define LOAD_FUNC(l, s) dlsym(l, s)
|
|
#define DL_CLOSE_FUNC(l) dlclose(l)
|
|
#endif
|
|
|
|
typedef enum cudaError_enum {
|
|
CUDA_SUCCESS = 0
|
|
} CUresult;
|
|
typedef int CUdevice;
|
|
typedef void* CUcontext;
|
|
|
|
typedef CUresult(CUDAAPI *PCUINIT)(unsigned int Flags);
|
|
typedef CUresult(CUDAAPI *PCUDEVICEGETCOUNT)(int *count);
|
|
typedef CUresult(CUDAAPI *PCUDEVICEGET)(CUdevice *device, int ordinal);
|
|
typedef CUresult(CUDAAPI *PCUDEVICEGETNAME)(char *name, int len, CUdevice dev);
|
|
typedef CUresult(CUDAAPI *PCUDEVICECOMPUTECAPABILITY)(int *major, int *minor, CUdevice dev);
|
|
typedef CUresult(CUDAAPI *PCUCTXCREATE)(CUcontext *pctx, unsigned int flags, CUdevice dev);
|
|
typedef CUresult(CUDAAPI *PCUCTXPOPCURRENT)(CUcontext *pctx);
|
|
typedef CUresult(CUDAAPI *PCUCTXDESTROY)(CUcontext ctx);
|
|
|
|
typedef NVENCSTATUS (NVENCAPI* PNVENCODEAPICREATEINSTANCE)(NV_ENCODE_API_FUNCTION_LIST *functionList);
|
|
|
|
typedef struct NvencInputSurface
|
|
{
|
|
NV_ENC_INPUT_PTR input_surface;
|
|
int width;
|
|
int height;
|
|
|
|
int lockCount;
|
|
|
|
NV_ENC_BUFFER_FORMAT format;
|
|
} NvencInputSurface;
|
|
|
|
typedef struct NvencOutputSurface
|
|
{
|
|
NV_ENC_OUTPUT_PTR output_surface;
|
|
int size;
|
|
|
|
NvencInputSurface* input_surface;
|
|
|
|
int busy;
|
|
} NvencOutputSurface;
|
|
|
|
typedef struct NvencData
|
|
{
|
|
union {
|
|
int64_t timestamp;
|
|
NvencOutputSurface *surface;
|
|
} u;
|
|
} NvencData;
|
|
|
|
typedef struct NvencDataList
|
|
{
|
|
NvencData* data;
|
|
|
|
uint32_t pos;
|
|
uint32_t count;
|
|
uint32_t size;
|
|
} NvencDataList;
|
|
|
|
typedef struct NvencDynLoadFunctions
|
|
{
|
|
PCUINIT cu_init;
|
|
PCUDEVICEGETCOUNT cu_device_get_count;
|
|
PCUDEVICEGET cu_device_get;
|
|
PCUDEVICEGETNAME cu_device_get_name;
|
|
PCUDEVICECOMPUTECAPABILITY cu_device_compute_capability;
|
|
PCUCTXCREATE cu_ctx_create;
|
|
PCUCTXPOPCURRENT cu_ctx_pop_current;
|
|
PCUCTXDESTROY cu_ctx_destroy;
|
|
|
|
NV_ENCODE_API_FUNCTION_LIST nvenc_funcs;
|
|
int nvenc_device_count;
|
|
CUdevice nvenc_devices[16];
|
|
|
|
#if defined(_WIN32)
|
|
HMODULE cuda_lib;
|
|
HMODULE nvenc_lib;
|
|
#else
|
|
void* cuda_lib;
|
|
void* nvenc_lib;
|
|
#endif
|
|
} NvencDynLoadFunctions;
|
|
|
|
typedef struct NvencValuePair
|
|
{
|
|
const char *str;
|
|
uint32_t num;
|
|
} NvencValuePair;
|
|
|
|
typedef struct NvencContext
|
|
{
|
|
AVClass *avclass;
|
|
|
|
NvencDynLoadFunctions nvenc_dload_funcs;
|
|
|
|
NV_ENC_INITIALIZE_PARAMS init_encode_params;
|
|
NV_ENC_CONFIG encode_config;
|
|
CUcontext cu_context;
|
|
|
|
int max_surface_count;
|
|
NvencInputSurface *input_surfaces;
|
|
NvencOutputSurface *output_surfaces;
|
|
|
|
NvencDataList output_surface_queue;
|
|
NvencDataList output_surface_ready_queue;
|
|
NvencDataList timestamp_list;
|
|
int64_t last_dts;
|
|
|
|
void *nvencoder;
|
|
|
|
char *preset;
|
|
char *profile;
|
|
char *level;
|
|
char *tier;
|
|
int cbr;
|
|
int twopass;
|
|
int gpu;
|
|
int buffer_delay;
|
|
} NvencContext;
|
|
|
|
static const NvencValuePair nvenc_h264_level_pairs[] = {
|
|
{ "auto", NV_ENC_LEVEL_AUTOSELECT },
|
|
{ "1" , NV_ENC_LEVEL_H264_1 },
|
|
{ "1.0" , NV_ENC_LEVEL_H264_1 },
|
|
{ "1b" , NV_ENC_LEVEL_H264_1b },
|
|
{ "1.0b", NV_ENC_LEVEL_H264_1b },
|
|
{ "1.1" , NV_ENC_LEVEL_H264_11 },
|
|
{ "1.2" , NV_ENC_LEVEL_H264_12 },
|
|
{ "1.3" , NV_ENC_LEVEL_H264_13 },
|
|
{ "2" , NV_ENC_LEVEL_H264_2 },
|
|
{ "2.0" , NV_ENC_LEVEL_H264_2 },
|
|
{ "2.1" , NV_ENC_LEVEL_H264_21 },
|
|
{ "2.2" , NV_ENC_LEVEL_H264_22 },
|
|
{ "3" , NV_ENC_LEVEL_H264_3 },
|
|
{ "3.0" , NV_ENC_LEVEL_H264_3 },
|
|
{ "3.1" , NV_ENC_LEVEL_H264_31 },
|
|
{ "3.2" , NV_ENC_LEVEL_H264_32 },
|
|
{ "4" , NV_ENC_LEVEL_H264_4 },
|
|
{ "4.0" , NV_ENC_LEVEL_H264_4 },
|
|
{ "4.1" , NV_ENC_LEVEL_H264_41 },
|
|
{ "4.2" , NV_ENC_LEVEL_H264_42 },
|
|
{ "5" , NV_ENC_LEVEL_H264_5 },
|
|
{ "5.0" , NV_ENC_LEVEL_H264_5 },
|
|
{ "5.1" , NV_ENC_LEVEL_H264_51 },
|
|
{ NULL }
|
|
};
|
|
|
|
static const NvencValuePair nvenc_hevc_level_pairs[] = {
|
|
{ "auto", NV_ENC_LEVEL_AUTOSELECT },
|
|
{ "1" , NV_ENC_LEVEL_HEVC_1 },
|
|
{ "1.0" , NV_ENC_LEVEL_HEVC_1 },
|
|
{ "2" , NV_ENC_LEVEL_HEVC_2 },
|
|
{ "2.0" , NV_ENC_LEVEL_HEVC_2 },
|
|
{ "2.1" , NV_ENC_LEVEL_HEVC_21 },
|
|
{ "3" , NV_ENC_LEVEL_HEVC_3 },
|
|
{ "3.0" , NV_ENC_LEVEL_HEVC_3 },
|
|
{ "3.1" , NV_ENC_LEVEL_HEVC_31 },
|
|
{ "4" , NV_ENC_LEVEL_HEVC_4 },
|
|
{ "4.0" , NV_ENC_LEVEL_HEVC_4 },
|
|
{ "4.1" , NV_ENC_LEVEL_HEVC_41 },
|
|
{ "5" , NV_ENC_LEVEL_HEVC_5 },
|
|
{ "5.0" , NV_ENC_LEVEL_HEVC_5 },
|
|
{ "5.1" , NV_ENC_LEVEL_HEVC_51 },
|
|
{ "5.2" , NV_ENC_LEVEL_HEVC_52 },
|
|
{ "6" , NV_ENC_LEVEL_HEVC_6 },
|
|
{ "6.0" , NV_ENC_LEVEL_HEVC_6 },
|
|
{ "6.1" , NV_ENC_LEVEL_HEVC_61 },
|
|
{ "6.2" , NV_ENC_LEVEL_HEVC_62 },
|
|
{ NULL }
|
|
};
|
|
|
|
static int input_string_to_uint32(AVCodecContext *avctx, const NvencValuePair *pair, const char *input, uint32_t *output)
|
|
{
|
|
for (; pair->str; ++pair) {
|
|
if (!strcmp(input, pair->str)) {
|
|
*output = pair->num;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
static NvencData* data_queue_dequeue(NvencDataList* queue)
|
|
{
|
|
uint32_t mask;
|
|
uint32_t read_pos;
|
|
|
|
av_assert0(queue);
|
|
av_assert0(queue->size);
|
|
av_assert0(queue->data);
|
|
|
|
if (!queue->count)
|
|
return NULL;
|
|
|
|
/* Size always is a multiple of two */
|
|
mask = queue->size - 1;
|
|
read_pos = (queue->pos - queue->count) & mask;
|
|
queue->count--;
|
|
|
|
return &queue->data[read_pos];
|
|
}
|
|
|
|
static int data_queue_enqueue(NvencDataList* queue, NvencData *data)
|
|
{
|
|
NvencDataList new_queue;
|
|
NvencData* tmp_data;
|
|
uint32_t mask;
|
|
|
|
if (!queue->size) {
|
|
/* size always has to be a multiple of two */
|
|
queue->size = 4;
|
|
queue->pos = 0;
|
|
queue->count = 0;
|
|
|
|
queue->data = av_malloc(queue->size * sizeof(*(queue->data)));
|
|
|
|
if (!queue->data) {
|
|
queue->size = 0;
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
}
|
|
|
|
if (queue->count == queue->size) {
|
|
new_queue.size = queue->size << 1;
|
|
new_queue.pos = 0;
|
|
new_queue.count = 0;
|
|
new_queue.data = av_malloc(new_queue.size * sizeof(*(queue->data)));
|
|
|
|
if (!new_queue.data)
|
|
return AVERROR(ENOMEM);
|
|
|
|
while (tmp_data = data_queue_dequeue(queue))
|
|
data_queue_enqueue(&new_queue, tmp_data);
|
|
|
|
av_free(queue->data);
|
|
*queue = new_queue;
|
|
}
|
|
|
|
mask = queue->size - 1;
|
|
|
|
queue->data[queue->pos] = *data;
|
|
queue->pos = (queue->pos + 1) & mask;
|
|
queue->count++;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int out_surf_queue_enqueue(NvencDataList* queue, NvencOutputSurface* surface)
|
|
{
|
|
NvencData data;
|
|
data.u.surface = surface;
|
|
|
|
return data_queue_enqueue(queue, &data);
|
|
}
|
|
|
|
static NvencOutputSurface* out_surf_queue_dequeue(NvencDataList* queue)
|
|
{
|
|
NvencData* res = data_queue_dequeue(queue);
|
|
|
|
if (!res)
|
|
return NULL;
|
|
|
|
return res->u.surface;
|
|
}
|
|
|
|
static int timestamp_queue_enqueue(NvencDataList* queue, int64_t timestamp)
|
|
{
|
|
NvencData data;
|
|
data.u.timestamp = timestamp;
|
|
|
|
return data_queue_enqueue(queue, &data);
|
|
}
|
|
|
|
static int64_t timestamp_queue_dequeue(NvencDataList* queue)
|
|
{
|
|
NvencData* res = data_queue_dequeue(queue);
|
|
|
|
if (!res)
|
|
return AV_NOPTS_VALUE;
|
|
|
|
return res->u.timestamp;
|
|
}
|
|
|
|
#define CHECK_LOAD_FUNC(t, f, s) \
|
|
do { \
|
|
(f) = (t)LOAD_FUNC(dl_fn->cuda_lib, s); \
|
|
if (!(f)) { \
|
|
av_log(avctx, AV_LOG_FATAL, "Failed loading %s from CUDA library\n", s); \
|
|
goto error; \
|
|
} \
|
|
} while (0)
|
|
|
|
static av_cold int nvenc_dyload_cuda(AVCodecContext *avctx)
|
|
{
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
|
|
if (dl_fn->cuda_lib)
|
|
return 1;
|
|
|
|
#if defined(_WIN32)
|
|
dl_fn->cuda_lib = LoadLibrary(TEXT("nvcuda.dll"));
|
|
#else
|
|
dl_fn->cuda_lib = dlopen("libcuda.so", RTLD_LAZY);
|
|
#endif
|
|
|
|
if (!dl_fn->cuda_lib) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed loading CUDA library\n");
|
|
goto error;
|
|
}
|
|
|
|
CHECK_LOAD_FUNC(PCUINIT, dl_fn->cu_init, "cuInit");
|
|
CHECK_LOAD_FUNC(PCUDEVICEGETCOUNT, dl_fn->cu_device_get_count, "cuDeviceGetCount");
|
|
CHECK_LOAD_FUNC(PCUDEVICEGET, dl_fn->cu_device_get, "cuDeviceGet");
|
|
CHECK_LOAD_FUNC(PCUDEVICEGETNAME, dl_fn->cu_device_get_name, "cuDeviceGetName");
|
|
CHECK_LOAD_FUNC(PCUDEVICECOMPUTECAPABILITY, dl_fn->cu_device_compute_capability, "cuDeviceComputeCapability");
|
|
CHECK_LOAD_FUNC(PCUCTXCREATE, dl_fn->cu_ctx_create, "cuCtxCreate_v2");
|
|
CHECK_LOAD_FUNC(PCUCTXPOPCURRENT, dl_fn->cu_ctx_pop_current, "cuCtxPopCurrent_v2");
|
|
CHECK_LOAD_FUNC(PCUCTXDESTROY, dl_fn->cu_ctx_destroy, "cuCtxDestroy_v2");
|
|
|
|
return 1;
|
|
|
|
error:
|
|
|
|
if (dl_fn->cuda_lib)
|
|
DL_CLOSE_FUNC(dl_fn->cuda_lib);
|
|
|
|
dl_fn->cuda_lib = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int check_cuda_errors(AVCodecContext *avctx, CUresult err, const char *func)
|
|
{
|
|
if (err != CUDA_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, ">> %s - failed with error code 0x%x\n", func, err);
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
#define check_cuda_errors(f) if (!check_cuda_errors(avctx, f, #f)) goto error
|
|
|
|
static av_cold int nvenc_check_cuda(AVCodecContext *avctx)
|
|
{
|
|
int device_count = 0;
|
|
CUdevice cu_device = 0;
|
|
char gpu_name[128];
|
|
int smminor = 0, smmajor = 0;
|
|
int i, smver, target_smver;
|
|
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
target_smver = avctx->pix_fmt == AV_PIX_FMT_YUV444P ? 0x52 : 0x30;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
target_smver = 0x52;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_FATAL, "Unknown codec name\n");
|
|
goto error;
|
|
}
|
|
|
|
if (!nvenc_dyload_cuda(avctx))
|
|
return 0;
|
|
|
|
if (dl_fn->nvenc_device_count > 0)
|
|
return 1;
|
|
|
|
check_cuda_errors(dl_fn->cu_init(0));
|
|
|
|
check_cuda_errors(dl_fn->cu_device_get_count(&device_count));
|
|
|
|
if (!device_count) {
|
|
av_log(avctx, AV_LOG_FATAL, "No CUDA capable devices found\n");
|
|
goto error;
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "%d CUDA capable devices found\n", device_count);
|
|
|
|
dl_fn->nvenc_device_count = 0;
|
|
|
|
for (i = 0; i < device_count; ++i) {
|
|
check_cuda_errors(dl_fn->cu_device_get(&cu_device, i));
|
|
check_cuda_errors(dl_fn->cu_device_get_name(gpu_name, sizeof(gpu_name), cu_device));
|
|
check_cuda_errors(dl_fn->cu_device_compute_capability(&smmajor, &smminor, cu_device));
|
|
|
|
smver = (smmajor << 4) | smminor;
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "[ GPU #%d - < %s > has Compute SM %d.%d, NVENC %s ]\n", i, gpu_name, smmajor, smminor, (smver >= target_smver) ? "Available" : "Not Available");
|
|
|
|
if (smver >= target_smver)
|
|
dl_fn->nvenc_devices[dl_fn->nvenc_device_count++] = cu_device;
|
|
}
|
|
|
|
if (!dl_fn->nvenc_device_count) {
|
|
av_log(avctx, AV_LOG_FATAL, "No NVENC capable devices found\n");
|
|
goto error;
|
|
}
|
|
|
|
return 1;
|
|
|
|
error:
|
|
|
|
dl_fn->nvenc_device_count = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold int nvenc_dyload_nvenc(AVCodecContext *avctx)
|
|
{
|
|
PNVENCODEAPICREATEINSTANCE nvEncodeAPICreateInstance = 0;
|
|
NVENCSTATUS nvstatus;
|
|
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
|
|
if (!nvenc_check_cuda(avctx))
|
|
return 0;
|
|
|
|
if (dl_fn->nvenc_lib)
|
|
return 1;
|
|
|
|
#if defined(_WIN32)
|
|
if (sizeof(void*) == 8) {
|
|
dl_fn->nvenc_lib = LoadLibrary(TEXT("nvEncodeAPI64.dll"));
|
|
} else {
|
|
dl_fn->nvenc_lib = LoadLibrary(TEXT("nvEncodeAPI.dll"));
|
|
}
|
|
#else
|
|
dl_fn->nvenc_lib = dlopen("libnvidia-encode.so.1", RTLD_LAZY);
|
|
#endif
|
|
|
|
if (!dl_fn->nvenc_lib) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed loading the nvenc library\n");
|
|
goto error;
|
|
}
|
|
|
|
nvEncodeAPICreateInstance = (PNVENCODEAPICREATEINSTANCE)LOAD_FUNC(dl_fn->nvenc_lib, "NvEncodeAPICreateInstance");
|
|
|
|
if (!nvEncodeAPICreateInstance) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed to load nvenc entrypoint\n");
|
|
goto error;
|
|
}
|
|
|
|
dl_fn->nvenc_funcs.version = NV_ENCODE_API_FUNCTION_LIST_VER;
|
|
|
|
nvstatus = nvEncodeAPICreateInstance(&dl_fn->nvenc_funcs);
|
|
|
|
if (nvstatus != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed to create nvenc instance\n");
|
|
goto error;
|
|
}
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "Nvenc initialized successfully\n");
|
|
|
|
return 1;
|
|
|
|
error:
|
|
if (dl_fn->nvenc_lib)
|
|
DL_CLOSE_FUNC(dl_fn->nvenc_lib);
|
|
|
|
dl_fn->nvenc_lib = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static av_cold void nvenc_unload_nvenc(AVCodecContext *avctx)
|
|
{
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
|
|
DL_CLOSE_FUNC(dl_fn->nvenc_lib);
|
|
dl_fn->nvenc_lib = NULL;
|
|
|
|
dl_fn->nvenc_device_count = 0;
|
|
|
|
DL_CLOSE_FUNC(dl_fn->cuda_lib);
|
|
dl_fn->cuda_lib = NULL;
|
|
|
|
dl_fn->cu_init = NULL;
|
|
dl_fn->cu_device_get_count = NULL;
|
|
dl_fn->cu_device_get = NULL;
|
|
dl_fn->cu_device_get_name = NULL;
|
|
dl_fn->cu_device_compute_capability = NULL;
|
|
dl_fn->cu_ctx_create = NULL;
|
|
dl_fn->cu_ctx_pop_current = NULL;
|
|
dl_fn->cu_ctx_destroy = NULL;
|
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "Nvenc unloaded\n");
|
|
}
|
|
|
|
static av_cold int nvenc_encode_init(AVCodecContext *avctx)
|
|
{
|
|
NV_ENC_OPEN_ENCODE_SESSION_EX_PARAMS encode_session_params = { 0 };
|
|
NV_ENC_PRESET_CONFIG preset_config = { 0 };
|
|
CUcontext cu_context_curr;
|
|
CUresult cu_res;
|
|
GUID encoder_preset = NV_ENC_PRESET_HQ_GUID;
|
|
GUID codec;
|
|
NVENCSTATUS nv_status = NV_ENC_SUCCESS;
|
|
AVCPBProperties *cpb_props;
|
|
int surfaceCount = 0;
|
|
int i, num_mbs;
|
|
int isLL = 0;
|
|
int lossless = 0;
|
|
int res = 0;
|
|
int dw, dh;
|
|
int qp_inter_p;
|
|
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
NV_ENCODE_API_FUNCTION_LIST *p_nvenc = &dl_fn->nvenc_funcs;
|
|
|
|
if (!nvenc_dyload_nvenc(avctx))
|
|
return AVERROR_EXTERNAL;
|
|
|
|
ctx->last_dts = AV_NOPTS_VALUE;
|
|
|
|
ctx->encode_config.version = NV_ENC_CONFIG_VER;
|
|
ctx->init_encode_params.version = NV_ENC_INITIALIZE_PARAMS_VER;
|
|
preset_config.version = NV_ENC_PRESET_CONFIG_VER;
|
|
preset_config.presetCfg.version = NV_ENC_CONFIG_VER;
|
|
encode_session_params.version = NV_ENC_OPEN_ENCODE_SESSION_EX_PARAMS_VER;
|
|
encode_session_params.apiVersion = NVENCAPI_VERSION;
|
|
|
|
if (ctx->gpu >= dl_fn->nvenc_device_count) {
|
|
av_log(avctx, AV_LOG_FATAL, "Requested GPU %d, but only %d GPUs are available!\n", ctx->gpu, dl_fn->nvenc_device_count);
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
|
|
ctx->cu_context = NULL;
|
|
cu_res = dl_fn->cu_ctx_create(&ctx->cu_context, 4, dl_fn->nvenc_devices[ctx->gpu]); // CU_CTX_SCHED_BLOCKING_SYNC=4, avoid CPU spins
|
|
|
|
if (cu_res != CUDA_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed creating CUDA context for NVENC: 0x%x\n", (int)cu_res);
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
cu_res = dl_fn->cu_ctx_pop_current(&cu_context_curr);
|
|
|
|
if (cu_res != CUDA_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed popping CUDA context: 0x%x\n", (int)cu_res);
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
encode_session_params.device = ctx->cu_context;
|
|
encode_session_params.deviceType = NV_ENC_DEVICE_TYPE_CUDA;
|
|
|
|
nv_status = p_nvenc->nvEncOpenEncodeSessionEx(&encode_session_params, &ctx->nvencoder);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
ctx->nvencoder = NULL;
|
|
av_log(avctx, AV_LOG_FATAL, "OpenEncodeSessionEx failed: 0x%x\n", (int)nv_status);
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
if (ctx->preset) {
|
|
if (!strcmp(ctx->preset, "slow")) {
|
|
encoder_preset = NV_ENC_PRESET_HQ_GUID;
|
|
ctx->twopass = 1;
|
|
} else if (!strcmp(ctx->preset, "medium")) {
|
|
encoder_preset = NV_ENC_PRESET_HQ_GUID;
|
|
ctx->twopass = 0;
|
|
} else if (!strcmp(ctx->preset, "fast")) {
|
|
encoder_preset = NV_ENC_PRESET_HP_GUID;
|
|
ctx->twopass = 0;
|
|
} else if (!strcmp(ctx->preset, "hq")) {
|
|
encoder_preset = NV_ENC_PRESET_HQ_GUID;
|
|
} else if (!strcmp(ctx->preset, "hp")) {
|
|
encoder_preset = NV_ENC_PRESET_HP_GUID;
|
|
} else if (!strcmp(ctx->preset, "bd")) {
|
|
encoder_preset = NV_ENC_PRESET_BD_GUID;
|
|
} else if (!strcmp(ctx->preset, "ll")) {
|
|
encoder_preset = NV_ENC_PRESET_LOW_LATENCY_DEFAULT_GUID;
|
|
isLL = 1;
|
|
} else if (!strcmp(ctx->preset, "llhp")) {
|
|
encoder_preset = NV_ENC_PRESET_LOW_LATENCY_HP_GUID;
|
|
isLL = 1;
|
|
} else if (!strcmp(ctx->preset, "llhq")) {
|
|
encoder_preset = NV_ENC_PRESET_LOW_LATENCY_HQ_GUID;
|
|
isLL = 1;
|
|
} else if (!strcmp(ctx->preset, "lossless")) {
|
|
encoder_preset = NV_ENC_PRESET_LOSSLESS_DEFAULT_GUID;
|
|
lossless = 1;
|
|
} else if (!strcmp(ctx->preset, "losslesshp")) {
|
|
encoder_preset = NV_ENC_PRESET_LOSSLESS_HP_GUID;
|
|
lossless = 1;
|
|
} else if (!strcmp(ctx->preset, "default")) {
|
|
encoder_preset = NV_ENC_PRESET_DEFAULT_GUID;
|
|
} else {
|
|
av_log(avctx, AV_LOG_FATAL, "Preset \"%s\" is unknown! Supported presets: slow, medium, high, hp, hq, bd, ll, llhp, llhq, lossless, losslesshp, default\n", ctx->preset);
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
if (ctx->twopass < 0) {
|
|
ctx->twopass = isLL;
|
|
}
|
|
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
codec = NV_ENC_CODEC_H264_GUID;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
codec = NV_ENC_CODEC_HEVC_GUID;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_ERROR, "Unknown codec name\n");
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
|
|
nv_status = p_nvenc->nvEncGetEncodePresetConfig(ctx->nvencoder, codec, encoder_preset, &preset_config);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "GetEncodePresetConfig failed: 0x%x\n", (int)nv_status);
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
ctx->init_encode_params.encodeGUID = codec;
|
|
ctx->init_encode_params.encodeHeight = avctx->height;
|
|
ctx->init_encode_params.encodeWidth = avctx->width;
|
|
|
|
if (avctx->sample_aspect_ratio.num && avctx->sample_aspect_ratio.den &&
|
|
(avctx->sample_aspect_ratio.num != 1 || avctx->sample_aspect_ratio.num != 1)) {
|
|
av_reduce(&dw, &dh,
|
|
avctx->width * avctx->sample_aspect_ratio.num,
|
|
avctx->height * avctx->sample_aspect_ratio.den,
|
|
1024 * 1024);
|
|
ctx->init_encode_params.darHeight = dh;
|
|
ctx->init_encode_params.darWidth = dw;
|
|
} else {
|
|
ctx->init_encode_params.darHeight = avctx->height;
|
|
ctx->init_encode_params.darWidth = avctx->width;
|
|
}
|
|
|
|
// De-compensate for hardware, dubiously, trying to compensate for
|
|
// playback at 704 pixel width.
|
|
if (avctx->width == 720 &&
|
|
(avctx->height == 480 || avctx->height == 576)) {
|
|
av_reduce(&dw, &dh,
|
|
ctx->init_encode_params.darWidth * 44,
|
|
ctx->init_encode_params.darHeight * 45,
|
|
1024 * 1024);
|
|
ctx->init_encode_params.darHeight = dh;
|
|
ctx->init_encode_params.darWidth = dw;
|
|
}
|
|
|
|
ctx->init_encode_params.frameRateNum = avctx->time_base.den;
|
|
ctx->init_encode_params.frameRateDen = avctx->time_base.num * avctx->ticks_per_frame;
|
|
|
|
num_mbs = ((avctx->width + 15) >> 4) * ((avctx->height + 15) >> 4);
|
|
ctx->max_surface_count = (num_mbs >= 8160) ? 32 : 48;
|
|
|
|
if (ctx->buffer_delay >= ctx->max_surface_count)
|
|
ctx->buffer_delay = ctx->max_surface_count - 1;
|
|
|
|
ctx->init_encode_params.enableEncodeAsync = 0;
|
|
ctx->init_encode_params.enablePTD = 1;
|
|
|
|
ctx->init_encode_params.presetGUID = encoder_preset;
|
|
|
|
ctx->init_encode_params.encodeConfig = &ctx->encode_config;
|
|
memcpy(&ctx->encode_config, &preset_config.presetCfg, sizeof(ctx->encode_config));
|
|
ctx->encode_config.version = NV_ENC_CONFIG_VER;
|
|
|
|
if (avctx->refs >= 0) {
|
|
/* 0 means "let the hardware decide" */
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
ctx->encode_config.encodeCodecConfig.h264Config.maxNumRefFrames = avctx->refs;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.maxNumRefFramesInDPB = avctx->refs;
|
|
break;
|
|
/* Earlier switch/case will return if unknown codec is passed. */
|
|
}
|
|
}
|
|
|
|
if (avctx->gop_size > 0) {
|
|
if (avctx->max_b_frames >= 0) {
|
|
/* 0 is intra-only, 1 is I/P only, 2 is one B Frame, 3 two B frames, and so on. */
|
|
ctx->encode_config.frameIntervalP = avctx->max_b_frames + 1;
|
|
}
|
|
|
|
ctx->encode_config.gopLength = avctx->gop_size;
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
ctx->encode_config.encodeCodecConfig.h264Config.idrPeriod = avctx->gop_size;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.idrPeriod = avctx->gop_size;
|
|
break;
|
|
/* Earlier switch/case will return if unknown codec is passed. */
|
|
}
|
|
} else if (avctx->gop_size == 0) {
|
|
ctx->encode_config.frameIntervalP = 0;
|
|
ctx->encode_config.gopLength = 1;
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
ctx->encode_config.encodeCodecConfig.h264Config.idrPeriod = 1;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.idrPeriod = 1;
|
|
break;
|
|
/* Earlier switch/case will return if unknown codec is passed. */
|
|
}
|
|
}
|
|
|
|
/* when there're b frames, set dts offset */
|
|
if (ctx->encode_config.frameIntervalP >= 2)
|
|
ctx->last_dts = -2;
|
|
|
|
if (avctx->bit_rate > 0) {
|
|
ctx->encode_config.rcParams.averageBitRate = avctx->bit_rate;
|
|
} else if (ctx->encode_config.rcParams.averageBitRate > 0) {
|
|
ctx->encode_config.rcParams.maxBitRate = ctx->encode_config.rcParams.averageBitRate;
|
|
}
|
|
|
|
if (avctx->rc_max_rate > 0)
|
|
ctx->encode_config.rcParams.maxBitRate = avctx->rc_max_rate;
|
|
|
|
if (lossless) {
|
|
if (avctx->codec->id == AV_CODEC_ID_H264)
|
|
ctx->encode_config.encodeCodecConfig.h264Config.qpPrimeYZeroTransformBypassFlag = 1;
|
|
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_CONSTQP;
|
|
ctx->encode_config.rcParams.constQP.qpInterB = 0;
|
|
ctx->encode_config.rcParams.constQP.qpInterP = 0;
|
|
ctx->encode_config.rcParams.constQP.qpIntra = 0;
|
|
|
|
avctx->qmin = -1;
|
|
avctx->qmax = -1;
|
|
} else if (ctx->cbr) {
|
|
if (!ctx->twopass) {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_CBR;
|
|
} else {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_2_PASS_QUALITY;
|
|
|
|
if (avctx->codec->id == AV_CODEC_ID_H264) {
|
|
ctx->encode_config.encodeCodecConfig.h264Config.adaptiveTransformMode = NV_ENC_H264_ADAPTIVE_TRANSFORM_ENABLE;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.fmoMode = NV_ENC_H264_FMO_DISABLE;
|
|
}
|
|
}
|
|
} else if (avctx->global_quality > 0) {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_CONSTQP;
|
|
ctx->encode_config.rcParams.constQP.qpInterB = avctx->global_quality;
|
|
ctx->encode_config.rcParams.constQP.qpInterP = avctx->global_quality;
|
|
ctx->encode_config.rcParams.constQP.qpIntra = avctx->global_quality;
|
|
|
|
avctx->qmin = -1;
|
|
avctx->qmax = -1;
|
|
} else {
|
|
if (avctx->qmin >= 0 && avctx->qmax >= 0) {
|
|
ctx->encode_config.rcParams.enableMinQP = 1;
|
|
ctx->encode_config.rcParams.enableMaxQP = 1;
|
|
|
|
ctx->encode_config.rcParams.minQP.qpInterB = avctx->qmin;
|
|
ctx->encode_config.rcParams.minQP.qpInterP = avctx->qmin;
|
|
ctx->encode_config.rcParams.minQP.qpIntra = avctx->qmin;
|
|
|
|
ctx->encode_config.rcParams.maxQP.qpInterB = avctx->qmax;
|
|
ctx->encode_config.rcParams.maxQP.qpInterP = avctx->qmax;
|
|
ctx->encode_config.rcParams.maxQP.qpIntra = avctx->qmax;
|
|
|
|
qp_inter_p = (avctx->qmax + 3 * avctx->qmin) / 4; // biased towards Qmin
|
|
|
|
if (ctx->twopass) {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_2_PASS_VBR;
|
|
if (avctx->codec->id == AV_CODEC_ID_H264) {
|
|
ctx->encode_config.encodeCodecConfig.h264Config.adaptiveTransformMode = NV_ENC_H264_ADAPTIVE_TRANSFORM_ENABLE;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.fmoMode = NV_ENC_H264_FMO_DISABLE;
|
|
}
|
|
} else {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_VBR_MINQP;
|
|
}
|
|
} else {
|
|
qp_inter_p = 26; // default to 26
|
|
|
|
if (ctx->twopass) {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_2_PASS_VBR;
|
|
} else {
|
|
ctx->encode_config.rcParams.rateControlMode = NV_ENC_PARAMS_RC_VBR;
|
|
}
|
|
}
|
|
|
|
ctx->encode_config.rcParams.enableInitialRCQP = 1;
|
|
ctx->encode_config.rcParams.initialRCQP.qpInterP = qp_inter_p;
|
|
|
|
if(avctx->i_quant_factor != 0.0 && avctx->b_quant_factor != 0.0) {
|
|
ctx->encode_config.rcParams.initialRCQP.qpIntra = av_clip(
|
|
qp_inter_p * fabs(avctx->i_quant_factor) + avctx->i_quant_offset, 0, 51);
|
|
ctx->encode_config.rcParams.initialRCQP.qpInterB = av_clip(
|
|
qp_inter_p * fabs(avctx->b_quant_factor) + avctx->b_quant_offset, 0, 51);
|
|
} else {
|
|
ctx->encode_config.rcParams.initialRCQP.qpIntra = qp_inter_p;
|
|
ctx->encode_config.rcParams.initialRCQP.qpInterB = qp_inter_p;
|
|
}
|
|
}
|
|
|
|
if (avctx->rc_buffer_size > 0) {
|
|
ctx->encode_config.rcParams.vbvBufferSize = avctx->rc_buffer_size;
|
|
} else if (ctx->encode_config.rcParams.averageBitRate > 0) {
|
|
ctx->encode_config.rcParams.vbvBufferSize = 2 * ctx->encode_config.rcParams.averageBitRate;
|
|
}
|
|
|
|
if (avctx->flags & AV_CODEC_FLAG_INTERLACED_DCT) {
|
|
ctx->encode_config.frameFieldMode = NV_ENC_PARAMS_FRAME_FIELD_MODE_FIELD;
|
|
} else {
|
|
ctx->encode_config.frameFieldMode = NV_ENC_PARAMS_FRAME_FIELD_MODE_FRAME;
|
|
}
|
|
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.colourDescriptionPresentFlag = 1;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.videoSignalTypePresentFlag = 1;
|
|
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.colourMatrix = avctx->colorspace;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.colourPrimaries = avctx->color_primaries;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.transferCharacteristics = avctx->color_trc;
|
|
|
|
ctx->encode_config.encodeCodecConfig.h264Config.h264VUIParameters.videoFullRangeFlag = avctx->color_range == AVCOL_RANGE_JPEG;
|
|
|
|
ctx->encode_config.encodeCodecConfig.h264Config.sliceMode = 3;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.sliceModeData = 1;
|
|
|
|
ctx->encode_config.encodeCodecConfig.h264Config.disableSPSPPS = (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) ? 1 : 0;
|
|
ctx->encode_config.encodeCodecConfig.h264Config.repeatSPSPPS = (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) ? 0 : 1;
|
|
|
|
if (!ctx->profile) {
|
|
switch (avctx->profile) {
|
|
case FF_PROFILE_H264_HIGH_444_PREDICTIVE:
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_444_GUID;
|
|
break;
|
|
case FF_PROFILE_H264_BASELINE:
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_BASELINE_GUID;
|
|
break;
|
|
case FF_PROFILE_H264_MAIN:
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_MAIN_GUID;
|
|
break;
|
|
case FF_PROFILE_H264_HIGH:
|
|
case FF_PROFILE_UNKNOWN:
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_GUID;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_WARNING, "Unsupported profile requested, falling back to high\n");
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_GUID;
|
|
break;
|
|
}
|
|
} else {
|
|
if (!strcmp(ctx->profile, "high")) {
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_GUID;
|
|
avctx->profile = FF_PROFILE_H264_HIGH;
|
|
} else if (!strcmp(ctx->profile, "main")) {
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_MAIN_GUID;
|
|
avctx->profile = FF_PROFILE_H264_MAIN;
|
|
} else if (!strcmp(ctx->profile, "baseline")) {
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_BASELINE_GUID;
|
|
avctx->profile = FF_PROFILE_H264_BASELINE;
|
|
} else if (!strcmp(ctx->profile, "high444p")) {
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_444_GUID;
|
|
avctx->profile = FF_PROFILE_H264_HIGH_444_PREDICTIVE;
|
|
} else {
|
|
av_log(avctx, AV_LOG_FATAL, "Profile \"%s\" is unknown! Supported profiles: high, main, baseline\n", ctx->profile);
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
// force setting profile as high444p if input is AV_PIX_FMT_YUV444P
|
|
if (avctx->pix_fmt == AV_PIX_FMT_YUV444P) {
|
|
ctx->encode_config.profileGUID = NV_ENC_H264_PROFILE_HIGH_444_GUID;
|
|
avctx->profile = FF_PROFILE_H264_HIGH_444_PREDICTIVE;
|
|
}
|
|
|
|
ctx->encode_config.encodeCodecConfig.h264Config.chromaFormatIDC = avctx->profile == FF_PROFILE_H264_HIGH_444_PREDICTIVE ? 3 : 1;
|
|
|
|
if (ctx->level) {
|
|
res = input_string_to_uint32(avctx, nvenc_h264_level_pairs, ctx->level, &ctx->encode_config.encodeCodecConfig.h264Config.level);
|
|
|
|
if (res) {
|
|
av_log(avctx, AV_LOG_FATAL, "Level \"%s\" is unknown! Supported levels: auto, 1, 1b, 1.1, 1.2, 1.3, 2, 2.1, 2.2, 3, 3.1, 3.2, 4, 4.1, 4.2, 5, 5.1\n", ctx->level);
|
|
goto error;
|
|
}
|
|
} else {
|
|
ctx->encode_config.encodeCodecConfig.h264Config.level = NV_ENC_LEVEL_AUTOSELECT;
|
|
}
|
|
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.sliceMode = 3;
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.sliceModeData = 1;
|
|
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.disableSPSPPS = (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) ? 1 : 0;
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.repeatSPSPPS = (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) ? 0 : 1;
|
|
|
|
/* No other profile is supported in the current SDK version 5 */
|
|
ctx->encode_config.profileGUID = NV_ENC_HEVC_PROFILE_MAIN_GUID;
|
|
avctx->profile = FF_PROFILE_HEVC_MAIN;
|
|
|
|
if (ctx->level) {
|
|
res = input_string_to_uint32(avctx, nvenc_hevc_level_pairs, ctx->level, &ctx->encode_config.encodeCodecConfig.hevcConfig.level);
|
|
|
|
if (res) {
|
|
av_log(avctx, AV_LOG_FATAL, "Level \"%s\" is unknown! Supported levels: auto, 1, 2, 2.1, 3, 3.1, 4, 4.1, 5, 5.1, 5.2, 6, 6.1, 6.2\n", ctx->level);
|
|
goto error;
|
|
}
|
|
} else {
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.level = NV_ENC_LEVEL_AUTOSELECT;
|
|
}
|
|
|
|
if (ctx->tier) {
|
|
if (!strcmp(ctx->tier, "main")) {
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.tier = NV_ENC_TIER_HEVC_MAIN;
|
|
} else if (!strcmp(ctx->tier, "high")) {
|
|
ctx->encode_config.encodeCodecConfig.hevcConfig.tier = NV_ENC_TIER_HEVC_HIGH;
|
|
} else {
|
|
av_log(avctx, AV_LOG_FATAL, "Tier \"%s\" is unknown! Supported tiers: main, high\n", ctx->tier);
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
}
|
|
|
|
break;
|
|
/* Earlier switch/case will return if unknown codec is passed. */
|
|
}
|
|
|
|
nv_status = p_nvenc->nvEncInitializeEncoder(ctx->nvencoder, &ctx->init_encode_params);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "InitializeEncoder failed: 0x%x\n", (int)nv_status);
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
ctx->input_surfaces = av_malloc(ctx->max_surface_count * sizeof(*ctx->input_surfaces));
|
|
|
|
if (!ctx->input_surfaces) {
|
|
res = AVERROR(ENOMEM);
|
|
goto error;
|
|
}
|
|
|
|
ctx->output_surfaces = av_malloc(ctx->max_surface_count * sizeof(*ctx->output_surfaces));
|
|
|
|
if (!ctx->output_surfaces) {
|
|
res = AVERROR(ENOMEM);
|
|
goto error;
|
|
}
|
|
|
|
for (surfaceCount = 0; surfaceCount < ctx->max_surface_count; ++surfaceCount) {
|
|
NV_ENC_CREATE_INPUT_BUFFER allocSurf = { 0 };
|
|
NV_ENC_CREATE_BITSTREAM_BUFFER allocOut = { 0 };
|
|
allocSurf.version = NV_ENC_CREATE_INPUT_BUFFER_VER;
|
|
allocOut.version = NV_ENC_CREATE_BITSTREAM_BUFFER_VER;
|
|
|
|
allocSurf.width = (avctx->width + 31) & ~31;
|
|
allocSurf.height = (avctx->height + 31) & ~31;
|
|
|
|
allocSurf.memoryHeap = NV_ENC_MEMORY_HEAP_SYSMEM_CACHED;
|
|
|
|
switch (avctx->pix_fmt) {
|
|
case AV_PIX_FMT_YUV420P:
|
|
allocSurf.bufferFmt = NV_ENC_BUFFER_FORMAT_YV12_PL;
|
|
break;
|
|
|
|
case AV_PIX_FMT_NV12:
|
|
allocSurf.bufferFmt = NV_ENC_BUFFER_FORMAT_NV12_PL;
|
|
break;
|
|
|
|
case AV_PIX_FMT_YUV444P:
|
|
allocSurf.bufferFmt = NV_ENC_BUFFER_FORMAT_YUV444_PL;
|
|
break;
|
|
|
|
default:
|
|
av_log(avctx, AV_LOG_FATAL, "Invalid input pixel format\n");
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
|
|
nv_status = p_nvenc->nvEncCreateInputBuffer(ctx->nvencoder, &allocSurf);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "CreateInputBuffer failed\n");
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
ctx->input_surfaces[surfaceCount].lockCount = 0;
|
|
ctx->input_surfaces[surfaceCount].input_surface = allocSurf.inputBuffer;
|
|
ctx->input_surfaces[surfaceCount].format = allocSurf.bufferFmt;
|
|
ctx->input_surfaces[surfaceCount].width = allocSurf.width;
|
|
ctx->input_surfaces[surfaceCount].height = allocSurf.height;
|
|
|
|
/* 1MB is large enough to hold most output frames. NVENC increases this automaticaly if it's not enough. */
|
|
allocOut.size = 1024 * 1024;
|
|
|
|
allocOut.memoryHeap = NV_ENC_MEMORY_HEAP_SYSMEM_CACHED;
|
|
|
|
nv_status = p_nvenc->nvEncCreateBitstreamBuffer(ctx->nvencoder, &allocOut);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "CreateBitstreamBuffer failed\n");
|
|
ctx->output_surfaces[surfaceCount++].output_surface = NULL;
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
ctx->output_surfaces[surfaceCount].output_surface = allocOut.bitstreamBuffer;
|
|
ctx->output_surfaces[surfaceCount].size = allocOut.size;
|
|
ctx->output_surfaces[surfaceCount].busy = 0;
|
|
}
|
|
|
|
if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
|
|
uint32_t outSize = 0;
|
|
char tmpHeader[256];
|
|
NV_ENC_SEQUENCE_PARAM_PAYLOAD payload = { 0 };
|
|
payload.version = NV_ENC_SEQUENCE_PARAM_PAYLOAD_VER;
|
|
|
|
payload.spsppsBuffer = tmpHeader;
|
|
payload.inBufferSize = sizeof(tmpHeader);
|
|
payload.outSPSPPSPayloadSize = &outSize;
|
|
|
|
nv_status = p_nvenc->nvEncGetSequenceParams(ctx->nvencoder, &payload);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "GetSequenceParams failed\n");
|
|
goto error;
|
|
}
|
|
|
|
avctx->extradata_size = outSize;
|
|
avctx->extradata = av_mallocz(outSize + AV_INPUT_BUFFER_PADDING_SIZE);
|
|
|
|
if (!avctx->extradata) {
|
|
res = AVERROR(ENOMEM);
|
|
goto error;
|
|
}
|
|
|
|
memcpy(avctx->extradata, tmpHeader, outSize);
|
|
}
|
|
|
|
if (ctx->encode_config.frameIntervalP > 1)
|
|
avctx->has_b_frames = 2;
|
|
|
|
if (ctx->encode_config.rcParams.averageBitRate > 0)
|
|
avctx->bit_rate = ctx->encode_config.rcParams.averageBitRate;
|
|
|
|
cpb_props = ff_add_cpb_side_data(avctx);
|
|
if (!cpb_props)
|
|
return AVERROR(ENOMEM);
|
|
cpb_props->max_bitrate = ctx->encode_config.rcParams.maxBitRate;
|
|
cpb_props->avg_bitrate = avctx->bit_rate;
|
|
cpb_props->buffer_size = ctx->encode_config.rcParams.vbvBufferSize;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
|
|
for (i = 0; i < surfaceCount; ++i) {
|
|
p_nvenc->nvEncDestroyInputBuffer(ctx->nvencoder, ctx->input_surfaces[i].input_surface);
|
|
if (ctx->output_surfaces[i].output_surface)
|
|
p_nvenc->nvEncDestroyBitstreamBuffer(ctx->nvencoder, ctx->output_surfaces[i].output_surface);
|
|
}
|
|
|
|
if (ctx->nvencoder)
|
|
p_nvenc->nvEncDestroyEncoder(ctx->nvencoder);
|
|
|
|
if (ctx->cu_context)
|
|
dl_fn->cu_ctx_destroy(ctx->cu_context);
|
|
|
|
nvenc_unload_nvenc(avctx);
|
|
|
|
ctx->nvencoder = NULL;
|
|
ctx->cu_context = NULL;
|
|
|
|
return res;
|
|
}
|
|
|
|
static av_cold int nvenc_encode_close(AVCodecContext *avctx)
|
|
{
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
NV_ENCODE_API_FUNCTION_LIST *p_nvenc = &dl_fn->nvenc_funcs;
|
|
int i;
|
|
|
|
av_freep(&ctx->timestamp_list.data);
|
|
av_freep(&ctx->output_surface_ready_queue.data);
|
|
av_freep(&ctx->output_surface_queue.data);
|
|
|
|
for (i = 0; i < ctx->max_surface_count; ++i) {
|
|
p_nvenc->nvEncDestroyInputBuffer(ctx->nvencoder, ctx->input_surfaces[i].input_surface);
|
|
p_nvenc->nvEncDestroyBitstreamBuffer(ctx->nvencoder, ctx->output_surfaces[i].output_surface);
|
|
}
|
|
ctx->max_surface_count = 0;
|
|
|
|
p_nvenc->nvEncDestroyEncoder(ctx->nvencoder);
|
|
ctx->nvencoder = NULL;
|
|
|
|
dl_fn->cu_ctx_destroy(ctx->cu_context);
|
|
ctx->cu_context = NULL;
|
|
|
|
nvenc_unload_nvenc(avctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int process_output_surface(AVCodecContext *avctx, AVPacket *pkt, NvencOutputSurface *tmpoutsurf)
|
|
{
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
NV_ENCODE_API_FUNCTION_LIST *p_nvenc = &dl_fn->nvenc_funcs;
|
|
|
|
uint32_t slice_mode_data;
|
|
uint32_t *slice_offsets;
|
|
NV_ENC_LOCK_BITSTREAM lock_params = { 0 };
|
|
NVENCSTATUS nv_status;
|
|
int res = 0;
|
|
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
slice_mode_data = ctx->encode_config.encodeCodecConfig.h264Config.sliceModeData;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
slice_mode_data = ctx->encode_config.encodeCodecConfig.hevcConfig.sliceModeData;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_ERROR, "Unknown codec name\n");
|
|
res = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
slice_offsets = av_mallocz(slice_mode_data * sizeof(*slice_offsets));
|
|
|
|
if (!slice_offsets)
|
|
return AVERROR(ENOMEM);
|
|
|
|
lock_params.version = NV_ENC_LOCK_BITSTREAM_VER;
|
|
|
|
lock_params.doNotWait = 0;
|
|
lock_params.outputBitstream = tmpoutsurf->output_surface;
|
|
lock_params.sliceOffsets = slice_offsets;
|
|
|
|
nv_status = p_nvenc->nvEncLockBitstream(ctx->nvencoder, &lock_params);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed locking bitstream buffer\n");
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
}
|
|
|
|
if (res = ff_alloc_packet2(avctx, pkt, lock_params.bitstreamSizeInBytes,0)) {
|
|
p_nvenc->nvEncUnlockBitstream(ctx->nvencoder, tmpoutsurf->output_surface);
|
|
goto error;
|
|
}
|
|
|
|
memcpy(pkt->data, lock_params.bitstreamBufferPtr, lock_params.bitstreamSizeInBytes);
|
|
|
|
nv_status = p_nvenc->nvEncUnlockBitstream(ctx->nvencoder, tmpoutsurf->output_surface);
|
|
if (nv_status != NV_ENC_SUCCESS)
|
|
av_log(avctx, AV_LOG_ERROR, "Failed unlocking bitstream buffer, expect the gates of mordor to open\n");
|
|
|
|
switch (lock_params.pictureType) {
|
|
case NV_ENC_PIC_TYPE_IDR:
|
|
pkt->flags |= AV_PKT_FLAG_KEY;
|
|
#if FF_API_CODED_FRAME
|
|
FF_DISABLE_DEPRECATION_WARNINGS
|
|
case NV_ENC_PIC_TYPE_I:
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I;
|
|
break;
|
|
case NV_ENC_PIC_TYPE_P:
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_P;
|
|
break;
|
|
case NV_ENC_PIC_TYPE_B:
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_B;
|
|
break;
|
|
case NV_ENC_PIC_TYPE_BI:
|
|
avctx->coded_frame->pict_type = AV_PICTURE_TYPE_BI;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_ERROR, "Unknown picture type encountered, expect the output to be broken.\n");
|
|
av_log(avctx, AV_LOG_ERROR, "Please report this error and include as much information on how to reproduce it as possible.\n");
|
|
res = AVERROR_EXTERNAL;
|
|
goto error;
|
|
FF_ENABLE_DEPRECATION_WARNINGS
|
|
#endif
|
|
}
|
|
|
|
pkt->pts = lock_params.outputTimeStamp;
|
|
pkt->dts = timestamp_queue_dequeue(&ctx->timestamp_list);
|
|
|
|
/* when there're b frame(s), set dts offset */
|
|
if (ctx->encode_config.frameIntervalP >= 2)
|
|
pkt->dts -= 1;
|
|
|
|
if (pkt->dts > pkt->pts)
|
|
pkt->dts = pkt->pts;
|
|
|
|
if (ctx->last_dts != AV_NOPTS_VALUE && pkt->dts <= ctx->last_dts)
|
|
pkt->dts = ctx->last_dts + 1;
|
|
|
|
ctx->last_dts = pkt->dts;
|
|
|
|
av_free(slice_offsets);
|
|
|
|
return 0;
|
|
|
|
error:
|
|
|
|
av_free(slice_offsets);
|
|
timestamp_queue_dequeue(&ctx->timestamp_list);
|
|
|
|
return res;
|
|
}
|
|
|
|
static int nvenc_encode_frame(AVCodecContext *avctx, AVPacket *pkt,
|
|
const AVFrame *frame, int *got_packet)
|
|
{
|
|
NVENCSTATUS nv_status;
|
|
NvencOutputSurface *tmpoutsurf;
|
|
int res, i = 0;
|
|
|
|
NvencContext *ctx = avctx->priv_data;
|
|
NvencDynLoadFunctions *dl_fn = &ctx->nvenc_dload_funcs;
|
|
NV_ENCODE_API_FUNCTION_LIST *p_nvenc = &dl_fn->nvenc_funcs;
|
|
|
|
NV_ENC_PIC_PARAMS pic_params = { 0 };
|
|
pic_params.version = NV_ENC_PIC_PARAMS_VER;
|
|
|
|
if (frame) {
|
|
NV_ENC_LOCK_INPUT_BUFFER lockBufferParams = { 0 };
|
|
NvencInputSurface *inSurf = NULL;
|
|
|
|
for (i = 0; i < ctx->max_surface_count; ++i) {
|
|
if (!ctx->input_surfaces[i].lockCount) {
|
|
inSurf = &ctx->input_surfaces[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
av_assert0(inSurf);
|
|
|
|
inSurf->lockCount = 1;
|
|
|
|
lockBufferParams.version = NV_ENC_LOCK_INPUT_BUFFER_VER;
|
|
lockBufferParams.inputBuffer = inSurf->input_surface;
|
|
|
|
nv_status = p_nvenc->nvEncLockInputBuffer(ctx->nvencoder, &lockBufferParams);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_ERROR, "Failed locking nvenc input buffer\n");
|
|
return 0;
|
|
}
|
|
|
|
if (avctx->pix_fmt == AV_PIX_FMT_YUV420P) {
|
|
uint8_t *buf = lockBufferParams.bufferDataPtr;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[0], frame->linesize[0],
|
|
avctx->width, avctx->height);
|
|
|
|
buf += inSurf->height * lockBufferParams.pitch;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch >> 1,
|
|
frame->data[2], frame->linesize[2],
|
|
avctx->width >> 1, avctx->height >> 1);
|
|
|
|
buf += (inSurf->height * lockBufferParams.pitch) >> 2;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch >> 1,
|
|
frame->data[1], frame->linesize[1],
|
|
avctx->width >> 1, avctx->height >> 1);
|
|
} else if (avctx->pix_fmt == AV_PIX_FMT_NV12) {
|
|
uint8_t *buf = lockBufferParams.bufferDataPtr;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[0], frame->linesize[0],
|
|
avctx->width, avctx->height);
|
|
|
|
buf += inSurf->height * lockBufferParams.pitch;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[1], frame->linesize[1],
|
|
avctx->width, avctx->height >> 1);
|
|
} else if (avctx->pix_fmt == AV_PIX_FMT_YUV444P) {
|
|
uint8_t *buf = lockBufferParams.bufferDataPtr;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[0], frame->linesize[0],
|
|
avctx->width, avctx->height);
|
|
|
|
buf += inSurf->height * lockBufferParams.pitch;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[1], frame->linesize[1],
|
|
avctx->width, avctx->height);
|
|
|
|
buf += inSurf->height * lockBufferParams.pitch;
|
|
|
|
av_image_copy_plane(buf, lockBufferParams.pitch,
|
|
frame->data[2], frame->linesize[2],
|
|
avctx->width, avctx->height);
|
|
} else {
|
|
av_log(avctx, AV_LOG_FATAL, "Invalid pixel format!\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
nv_status = p_nvenc->nvEncUnlockInputBuffer(ctx->nvencoder, inSurf->input_surface);
|
|
if (nv_status != NV_ENC_SUCCESS) {
|
|
av_log(avctx, AV_LOG_FATAL, "Failed unlocking input buffer!\n");
|
|
return AVERROR_EXTERNAL;
|
|
}
|
|
|
|
for (i = 0; i < ctx->max_surface_count; ++i)
|
|
if (!ctx->output_surfaces[i].busy)
|
|
break;
|
|
|
|
if (i == ctx->max_surface_count) {
|
|
inSurf->lockCount = 0;
|
|
av_log(avctx, AV_LOG_FATAL, "No free output surface found!\n");
|
|
return AVERROR_EXTERNAL;
|
|
}
|
|
|
|
ctx->output_surfaces[i].input_surface = inSurf;
|
|
|
|
pic_params.inputBuffer = inSurf->input_surface;
|
|
pic_params.bufferFmt = inSurf->format;
|
|
pic_params.inputWidth = avctx->width;
|
|
pic_params.inputHeight = avctx->height;
|
|
pic_params.outputBitstream = ctx->output_surfaces[i].output_surface;
|
|
pic_params.completionEvent = 0;
|
|
|
|
if (avctx->flags & AV_CODEC_FLAG_INTERLACED_DCT) {
|
|
if (frame->top_field_first) {
|
|
pic_params.pictureStruct = NV_ENC_PIC_STRUCT_FIELD_TOP_BOTTOM;
|
|
} else {
|
|
pic_params.pictureStruct = NV_ENC_PIC_STRUCT_FIELD_BOTTOM_TOP;
|
|
}
|
|
} else {
|
|
pic_params.pictureStruct = NV_ENC_PIC_STRUCT_FRAME;
|
|
}
|
|
|
|
pic_params.encodePicFlags = 0;
|
|
pic_params.inputTimeStamp = frame->pts;
|
|
pic_params.inputDuration = 0;
|
|
switch (avctx->codec->id) {
|
|
case AV_CODEC_ID_H264:
|
|
pic_params.codecPicParams.h264PicParams.sliceMode = ctx->encode_config.encodeCodecConfig.h264Config.sliceMode;
|
|
pic_params.codecPicParams.h264PicParams.sliceModeData = ctx->encode_config.encodeCodecConfig.h264Config.sliceModeData;
|
|
break;
|
|
case AV_CODEC_ID_H265:
|
|
pic_params.codecPicParams.hevcPicParams.sliceMode = ctx->encode_config.encodeCodecConfig.hevcConfig.sliceMode;
|
|
pic_params.codecPicParams.hevcPicParams.sliceModeData = ctx->encode_config.encodeCodecConfig.hevcConfig.sliceModeData;
|
|
break;
|
|
default:
|
|
av_log(avctx, AV_LOG_ERROR, "Unknown codec name\n");
|
|
return AVERROR(EINVAL);
|
|
}
|
|
|
|
res = timestamp_queue_enqueue(&ctx->timestamp_list, frame->pts);
|
|
|
|
if (res)
|
|
return res;
|
|
} else {
|
|
pic_params.encodePicFlags = NV_ENC_PIC_FLAG_EOS;
|
|
}
|
|
|
|
nv_status = p_nvenc->nvEncEncodePicture(ctx->nvencoder, &pic_params);
|
|
|
|
if (frame && nv_status == NV_ENC_ERR_NEED_MORE_INPUT) {
|
|
res = out_surf_queue_enqueue(&ctx->output_surface_queue, &ctx->output_surfaces[i]);
|
|
|
|
if (res)
|
|
return res;
|
|
|
|
ctx->output_surfaces[i].busy = 1;
|
|
}
|
|
|
|
if (nv_status != NV_ENC_SUCCESS && nv_status != NV_ENC_ERR_NEED_MORE_INPUT) {
|
|
av_log(avctx, AV_LOG_ERROR, "EncodePicture failed!\n");
|
|
return AVERROR_EXTERNAL;
|
|
}
|
|
|
|
if (nv_status != NV_ENC_ERR_NEED_MORE_INPUT) {
|
|
while (ctx->output_surface_queue.count) {
|
|
tmpoutsurf = out_surf_queue_dequeue(&ctx->output_surface_queue);
|
|
res = out_surf_queue_enqueue(&ctx->output_surface_ready_queue, tmpoutsurf);
|
|
|
|
if (res)
|
|
return res;
|
|
}
|
|
|
|
if (frame) {
|
|
res = out_surf_queue_enqueue(&ctx->output_surface_ready_queue, &ctx->output_surfaces[i]);
|
|
|
|
if (res)
|
|
return res;
|
|
|
|
ctx->output_surfaces[i].busy = 1;
|
|
}
|
|
}
|
|
|
|
if (ctx->output_surface_ready_queue.count && (!frame || ctx->output_surface_ready_queue.count + ctx->output_surface_queue.count >= ctx->buffer_delay)) {
|
|
tmpoutsurf = out_surf_queue_dequeue(&ctx->output_surface_ready_queue);
|
|
|
|
res = process_output_surface(avctx, pkt, tmpoutsurf);
|
|
|
|
if (res)
|
|
return res;
|
|
|
|
tmpoutsurf->busy = 0;
|
|
av_assert0(tmpoutsurf->input_surface->lockCount);
|
|
tmpoutsurf->input_surface->lockCount--;
|
|
|
|
*got_packet = 1;
|
|
} else {
|
|
*got_packet = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const enum AVPixelFormat pix_fmts_nvenc[] = {
|
|
AV_PIX_FMT_YUV420P,
|
|
AV_PIX_FMT_NV12,
|
|
AV_PIX_FMT_YUV444P,
|
|
AV_PIX_FMT_NONE
|
|
};
|
|
|
|
#define OFFSET(x) offsetof(NvencContext, x)
|
|
#define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
|
|
static const AVOption options[] = {
|
|
{ "preset", "Set the encoding preset (one of slow = hq 2pass, medium = hq, fast = hp, hq, hp, bd, ll, llhq, llhp, default)", OFFSET(preset), AV_OPT_TYPE_STRING, { .str = "medium" }, 0, 0, VE },
|
|
{ "profile", "Set the encoding profile (high, main, baseline or high444p)", OFFSET(profile), AV_OPT_TYPE_STRING, { .str = "main" }, 0, 0, VE },
|
|
{ "level", "Set the encoding level restriction (auto, 1.0, 1.0b, 1.1, 1.2, ..., 4.2, 5.0, 5.1)", OFFSET(level), AV_OPT_TYPE_STRING, { .str = "auto" }, 0, 0, VE },
|
|
{ "tier", "Set the encoding tier (main or high)", OFFSET(tier), AV_OPT_TYPE_STRING, { .str = "main" }, 0, 0, VE },
|
|
{ "cbr", "Use cbr encoding mode", OFFSET(cbr), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VE },
|
|
{ "2pass", "Use 2pass encoding mode", OFFSET(twopass), AV_OPT_TYPE_BOOL, { .i64 = -1 }, -1, 1, VE },
|
|
{ "gpu", "Selects which NVENC capable GPU to use. First GPU is 0, second is 1, and so on.", OFFSET(gpu), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, VE },
|
|
{ "delay", "Delays frame output by the given amount of frames.", OFFSET(buffer_delay), AV_OPT_TYPE_INT, { .i64 = INT_MAX }, 0, INT_MAX, VE },
|
|
{ NULL }
|
|
};
|
|
|
|
static const AVCodecDefault nvenc_defaults[] = {
|
|
{ "b", "2M" },
|
|
{ "qmin", "-1" },
|
|
{ "qmax", "-1" },
|
|
{ "qdiff", "-1" },
|
|
{ "qblur", "-1" },
|
|
{ "qcomp", "-1" },
|
|
{ "g", "250" },
|
|
{ "bf", "0" },
|
|
{ NULL },
|
|
};
|
|
|
|
#if CONFIG_NVENC_ENCODER
|
|
static const AVClass nvenc_class = {
|
|
.class_name = "nvenc",
|
|
.item_name = av_default_item_name,
|
|
.option = options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
AVCodec ff_nvenc_encoder = {
|
|
.name = "nvenc",
|
|
.long_name = NULL_IF_CONFIG_SMALL("NVIDIA NVENC h264 encoder"),
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.id = AV_CODEC_ID_H264,
|
|
.priv_data_size = sizeof(NvencContext),
|
|
.init = nvenc_encode_init,
|
|
.encode2 = nvenc_encode_frame,
|
|
.close = nvenc_encode_close,
|
|
.capabilities = AV_CODEC_CAP_DELAY,
|
|
.priv_class = &nvenc_class,
|
|
.defaults = nvenc_defaults,
|
|
.pix_fmts = pix_fmts_nvenc,
|
|
};
|
|
#endif
|
|
|
|
/* Add an alias for nvenc_h264 */
|
|
#if CONFIG_NVENC_H264_ENCODER
|
|
static const AVClass nvenc_h264_class = {
|
|
.class_name = "nvenc_h264",
|
|
.item_name = av_default_item_name,
|
|
.option = options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
AVCodec ff_nvenc_h264_encoder = {
|
|
.name = "nvenc_h264",
|
|
.long_name = NULL_IF_CONFIG_SMALL("NVIDIA NVENC h264 encoder"),
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.id = AV_CODEC_ID_H264,
|
|
.priv_data_size = sizeof(NvencContext),
|
|
.init = nvenc_encode_init,
|
|
.encode2 = nvenc_encode_frame,
|
|
.close = nvenc_encode_close,
|
|
.capabilities = AV_CODEC_CAP_DELAY,
|
|
.priv_class = &nvenc_h264_class,
|
|
.defaults = nvenc_defaults,
|
|
.pix_fmts = pix_fmts_nvenc,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_NVENC_HEVC_ENCODER
|
|
static const AVClass nvenc_hevc_class = {
|
|
.class_name = "nvenc_hevc",
|
|
.item_name = av_default_item_name,
|
|
.option = options,
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
};
|
|
|
|
AVCodec ff_nvenc_hevc_encoder = {
|
|
.name = "nvenc_hevc",
|
|
.long_name = NULL_IF_CONFIG_SMALL("NVIDIA NVENC hevc encoder"),
|
|
.type = AVMEDIA_TYPE_VIDEO,
|
|
.id = AV_CODEC_ID_H265,
|
|
.priv_data_size = sizeof(NvencContext),
|
|
.init = nvenc_encode_init,
|
|
.encode2 = nvenc_encode_frame,
|
|
.close = nvenc_encode_close,
|
|
.capabilities = AV_CODEC_CAP_DELAY,
|
|
.priv_class = &nvenc_hevc_class,
|
|
.defaults = nvenc_defaults,
|
|
.pix_fmts = pix_fmts_nvenc,
|
|
};
|
|
#endif
|