1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-12-07 11:13:41 +02:00
FFmpeg/libavdevice/decklink_enc.cpp

959 lines
32 KiB
C++
Raw Normal View History

/*
* Blackmagic DeckLink output
* Copyright (c) 2013-2014 Ramiro Polla
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include <atomic>
using std::atomic;
/* Include internal.h first to avoid conflict between winsock.h (used by
* DeckLink headers) and winsock2.h (used by libavformat) in MSVC++ builds */
extern "C" {
#include "libavformat/internal.h"
}
#include <DeckLinkAPI.h>
extern "C" {
#include "libavformat/avformat.h"
#include "libavcodec/bytestream.h"
#include "libavutil/internal.h"
#include "libavutil/imgutils.h"
#include "avdevice.h"
}
#include "decklink_common.h"
#include "decklink_enc.h"
#if CONFIG_LIBKLVANC
#include "libklvanc/vanc.h"
#include "libklvanc/vanc-lines.h"
#include "libklvanc/pixels.h"
#endif
/* DeckLink callback class declaration */
class decklink_frame : public IDeckLinkVideoFrame
{
public:
decklink_frame(struct decklink_ctx *ctx, AVFrame *avframe, AVCodecID codec_id, int height, int width) :
_ctx(ctx), _avframe(avframe), _avpacket(NULL), _codec_id(codec_id), _ancillary(NULL), _height(height), _width(width), _refs(1) { }
decklink_frame(struct decklink_ctx *ctx, AVPacket *avpacket, AVCodecID codec_id, int height, int width) :
_ctx(ctx), _avframe(NULL), _avpacket(avpacket), _codec_id(codec_id), _ancillary(NULL), _height(height), _width(width), _refs(1) { }
virtual long STDMETHODCALLTYPE GetWidth (void) { return _width; }
virtual long STDMETHODCALLTYPE GetHeight (void) { return _height; }
virtual long STDMETHODCALLTYPE GetRowBytes (void)
{
if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME)
return _avframe->linesize[0] < 0 ? -_avframe->linesize[0] : _avframe->linesize[0];
else
return ((GetWidth() + 47) / 48) * 128;
}
virtual BMDPixelFormat STDMETHODCALLTYPE GetPixelFormat(void)
{
if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME)
return bmdFormat8BitYUV;
else
return bmdFormat10BitYUV;
}
virtual BMDFrameFlags STDMETHODCALLTYPE GetFlags (void)
{
if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME)
return _avframe->linesize[0] < 0 ? bmdFrameFlagFlipVertical : bmdFrameFlagDefault;
else
return bmdFrameFlagDefault;
}
virtual HRESULT STDMETHODCALLTYPE GetBytes (void **buffer)
{
if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) {
if (_avframe->linesize[0] < 0)
*buffer = (void *)(_avframe->data[0] + _avframe->linesize[0] * (_avframe->height - 1));
else
*buffer = (void *)(_avframe->data[0]);
} else {
*buffer = (void *)(_avpacket->data);
}
return S_OK;
}
virtual HRESULT STDMETHODCALLTYPE GetTimecode (BMDTimecodeFormat format, IDeckLinkTimecode **timecode) { return S_FALSE; }
virtual HRESULT STDMETHODCALLTYPE GetAncillaryData(IDeckLinkVideoFrameAncillary **ancillary)
{
*ancillary = _ancillary;
if (_ancillary) {
_ancillary->AddRef();
return S_OK;
} else {
return S_FALSE;
}
}
virtual HRESULT STDMETHODCALLTYPE SetAncillaryData(IDeckLinkVideoFrameAncillary *ancillary)
{
if (_ancillary)
_ancillary->Release();
_ancillary = ancillary;
_ancillary->AddRef();
return S_OK;
}
virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; }
virtual ULONG STDMETHODCALLTYPE AddRef(void) { return ++_refs; }
virtual ULONG STDMETHODCALLTYPE Release(void)
{
int ret = --_refs;
if (!ret) {
av_frame_free(&_avframe);
av_packet_free(&_avpacket);
if (_ancillary)
_ancillary->Release();
delete this;
}
return ret;
}
struct decklink_ctx *_ctx;
AVFrame *_avframe;
AVPacket *_avpacket;
AVCodecID _codec_id;
IDeckLinkVideoFrameAncillary *_ancillary;
int _height;
int _width;
private:
std::atomic<int> _refs;
};
class decklink_output_callback : public IDeckLinkVideoOutputCallback
{
public:
virtual HRESULT STDMETHODCALLTYPE ScheduledFrameCompleted(IDeckLinkVideoFrame *_frame, BMDOutputFrameCompletionResult result)
{
decklink_frame *frame = static_cast<decklink_frame *>(_frame);
struct decklink_ctx *ctx = frame->_ctx;
if (frame->_avframe)
av_frame_unref(frame->_avframe);
if (frame->_avpacket)
av_packet_unref(frame->_avpacket);
avdevice/decklink: remove pthread dependency Purpose: avdevice/decklink: Removed pthread dependency by replacing semaphore used in code appropriately. Doing so makes it easier to build ffmpeg using Visual C++ on Windows. This is a contination of Kyle Schwarz's "avdevice/decklink: Remove pthread dependency" patch that is available at https://patchwork.ffmpeg.org/patch/2654/ . This patch wasn't accepted, and as far as I can tell, there was no follow-up after it was rejected. Notes: Used Visual Studio 2015 (with update 3) for this. Comments: -- configure: Eliminated pthreads dependency for decklink_indev_deps and decklink_outdev_deps and replaced with threads dependency -- libavdevice/decklink_common.cpp / .h: a) Eliminated semaphore and replaced with a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). b) Removed include of pthread.h and semaphore.h and now using libavutil/thread.h instead. -- libavdevice/decklink_dec.cpp: Eliminated include of pthread.h and semaphore.h. -- libavdevice/decklink_enc.cpp: a) Eliminated include of pthread.h and semaphore.h. b) Replaced use of semaphore with the equivalent using a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). In theory, libavutil/thread.h and the associated code could have been modified instead to add cross-platform implementations of the sem_ functions, but an inspection of the ffmpeg source base indicates that there are only two cases in which semaphores are used (including this one that was replaced), so it was deemed to not be worth the effort. Signed-off-by: Marton Balint <cus@passwd.hu>
2017-04-15 08:41:46 +02:00
pthread_mutex_lock(&ctx->mutex);
ctx->frames_buffer_available_spots++;
pthread_cond_broadcast(&ctx->cond);
pthread_mutex_unlock(&ctx->mutex);
return S_OK;
}
virtual HRESULT STDMETHODCALLTYPE ScheduledPlaybackHasStopped(void) { return S_OK; }
virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; }
virtual ULONG STDMETHODCALLTYPE AddRef(void) { return 1; }
virtual ULONG STDMETHODCALLTYPE Release(void) { return 1; }
};
static int decklink_setup_video(AVFormatContext *avctx, AVStream *st)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
AVCodecParameters *c = st->codecpar;
if (ctx->video) {
av_log(avctx, AV_LOG_ERROR, "Only one video stream is supported!\n");
return -1;
}
if (c->codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) {
if (c->format != AV_PIX_FMT_UYVY422) {
av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format!"
" Only AV_PIX_FMT_UYVY422 is supported.\n");
return -1;
}
ctx->raw_format = bmdFormat8BitYUV;
} else if (c->codec_id != AV_CODEC_ID_V210) {
av_log(avctx, AV_LOG_ERROR, "Unsupported codec type!"
" Only V210 and wrapped frame with AV_PIX_FMT_UYVY422 are supported.\n");
return -1;
} else {
ctx->raw_format = bmdFormat10BitYUV;
}
if (ff_decklink_set_configs(avctx, DIRECTION_OUT) < 0) {
av_log(avctx, AV_LOG_ERROR, "Could not set output configuration\n");
return -1;
}
if (ff_decklink_set_format(avctx, c->width, c->height,
st->time_base.num, st->time_base.den, c->field_order)) {
av_log(avctx, AV_LOG_ERROR, "Unsupported video size, framerate or field order!"
" Check available formats with -list_formats 1.\n");
return -1;
}
if (ctx->supports_vanc && ctx->dlo->EnableVideoOutput(ctx->bmd_mode, bmdVideoOutputVANC) != S_OK) {
av_log(avctx, AV_LOG_WARNING, "Could not enable video output with VANC! Trying without...\n");
ctx->supports_vanc = 0;
}
if (!ctx->supports_vanc && ctx->dlo->EnableVideoOutput(ctx->bmd_mode, bmdVideoOutputFlagDefault) != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not enable video output!\n");
return -1;
}
/* Set callback. */
ctx->output_callback = new decklink_output_callback();
ctx->dlo->SetScheduledFrameCompletionCallback(ctx->output_callback);
ctx->frames_preroll = st->time_base.den * ctx->preroll;
if (st->time_base.den > 1000)
ctx->frames_preroll /= 1000;
/* Buffer twice as many frames as the preroll. */
ctx->frames_buffer = ctx->frames_preroll * 2;
ctx->frames_buffer = FFMIN(ctx->frames_buffer, 60);
avdevice/decklink: remove pthread dependency Purpose: avdevice/decklink: Removed pthread dependency by replacing semaphore used in code appropriately. Doing so makes it easier to build ffmpeg using Visual C++ on Windows. This is a contination of Kyle Schwarz's "avdevice/decklink: Remove pthread dependency" patch that is available at https://patchwork.ffmpeg.org/patch/2654/ . This patch wasn't accepted, and as far as I can tell, there was no follow-up after it was rejected. Notes: Used Visual Studio 2015 (with update 3) for this. Comments: -- configure: Eliminated pthreads dependency for decklink_indev_deps and decklink_outdev_deps and replaced with threads dependency -- libavdevice/decklink_common.cpp / .h: a) Eliminated semaphore and replaced with a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). b) Removed include of pthread.h and semaphore.h and now using libavutil/thread.h instead. -- libavdevice/decklink_dec.cpp: Eliminated include of pthread.h and semaphore.h. -- libavdevice/decklink_enc.cpp: a) Eliminated include of pthread.h and semaphore.h. b) Replaced use of semaphore with the equivalent using a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). In theory, libavutil/thread.h and the associated code could have been modified instead to add cross-platform implementations of the sem_ functions, but an inspection of the ffmpeg source base indicates that there are only two cases in which semaphores are used (including this one that was replaced), so it was deemed to not be worth the effort. Signed-off-by: Marton Balint <cus@passwd.hu>
2017-04-15 08:41:46 +02:00
pthread_mutex_init(&ctx->mutex, NULL);
pthread_cond_init(&ctx->cond, NULL);
ctx->frames_buffer_available_spots = ctx->frames_buffer;
av_log(avctx, AV_LOG_DEBUG, "output: %s, preroll: %d, frames buffer size: %d\n",
avctx->url, ctx->frames_preroll, ctx->frames_buffer);
/* The device expects the framerate to be fixed. */
avpriv_set_pts_info(st, 64, st->time_base.num, st->time_base.den);
ctx->video = 1;
return 0;
}
static int decklink_setup_audio(AVFormatContext *avctx, AVStream *st)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
AVCodecParameters *c = st->codecpar;
if (ctx->audio) {
av_log(avctx, AV_LOG_ERROR, "Only one audio stream is supported!\n");
return -1;
}
if (c->codec_id == AV_CODEC_ID_AC3) {
/* Regardless of the number of channels in the codec, we're only
using 2 SDI audio channels at 48000Hz */
ctx->channels = 2;
} else if (c->codec_id == AV_CODEC_ID_PCM_S16LE) {
if (c->sample_rate != 48000) {
av_log(avctx, AV_LOG_ERROR, "Unsupported sample rate!"
" Only 48kHz is supported.\n");
return -1;
}
if (c->ch_layout.nb_channels != 2 && c->ch_layout.nb_channels != 8 && c->ch_layout.nb_channels != 16) {
av_log(avctx, AV_LOG_ERROR, "Unsupported number of channels!"
" Only 2, 8 or 16 channels are supported.\n");
return -1;
}
ctx->channels = c->ch_layout.nb_channels;
} else {
av_log(avctx, AV_LOG_ERROR, "Unsupported codec specified!"
" Only PCM_S16LE and AC-3 are supported.\n");
return -1;
}
if (ctx->dlo->EnableAudioOutput(bmdAudioSampleRate48kHz,
bmdAudioSampleType16bitInteger,
ctx->channels,
bmdAudioOutputStreamTimestamped) != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not enable audio output!\n");
return -1;
}
if (ctx->dlo->BeginAudioPreroll() != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not begin audio preroll!\n");
return -1;
}
/* The device expects the sample rate to be fixed. */
avpriv_set_pts_info(st, 64, 1, 48000);
ctx->audio = 1;
return 0;
}
/* Wrap the AC-3 packet into an S337 payload that is in S16LE format which can be easily
injected into the PCM stream. Note: despite the function name, only AC-3 is implemented */
static int create_s337_payload(AVPacket *pkt, uint8_t **outbuf, int *outsize)
{
/* Note: if the packet size is not divisible by four, we need to make the actual
payload larger to ensure it ends on an two channel S16LE boundary */
int payload_size = FFALIGN(pkt->size, 4) + 8;
uint16_t bitcount = pkt->size * 8;
uint8_t *s337_payload;
PutByteContext pb;
/* Sanity check: According to SMPTE ST 340:2015 Sec 4.1, the AC-3 sync frame will
exactly match the 1536 samples of baseband (PCM) audio that it represents. */
if (pkt->size > 1536)
return AVERROR(EINVAL);
/* Encapsulate AC3 syncframe into SMPTE 337 packet */
s337_payload = (uint8_t *) av_malloc(payload_size);
if (s337_payload == NULL)
return AVERROR(ENOMEM);
bytestream2_init_writer(&pb, s337_payload, payload_size);
bytestream2_put_le16u(&pb, 0xf872); /* Sync word 1 */
bytestream2_put_le16u(&pb, 0x4e1f); /* Sync word 1 */
bytestream2_put_le16u(&pb, 0x0001); /* Burst Info, including data type (1=ac3) */
bytestream2_put_le16u(&pb, bitcount); /* Length code */
for (int i = 0; i < (pkt->size - 1); i += 2)
bytestream2_put_le16u(&pb, (pkt->data[i] << 8) | pkt->data[i+1]);
/* Ensure final payload is aligned on 4-byte boundary */
if (pkt->size & 1)
bytestream2_put_le16u(&pb, pkt->data[pkt->size - 1] << 8);
if ((pkt->size & 3) == 1 || (pkt->size & 3) == 2)
bytestream2_put_le16u(&pb, 0);
*outsize = payload_size;
*outbuf = s337_payload;
return 0;
}
static int decklink_setup_subtitle(AVFormatContext *avctx, AVStream *st)
{
int ret = -1;
switch(st->codecpar->codec_id) {
#if CONFIG_LIBKLVANC
case AV_CODEC_ID_EIA_608:
/* No special setup required */
ret = 0;
break;
#endif
default:
av_log(avctx, AV_LOG_ERROR, "Unsupported subtitle codec specified\n");
break;
}
return ret;
}
static int decklink_setup_data(AVFormatContext *avctx, AVStream *st)
{
int ret = -1;
switch(st->codecpar->codec_id) {
#if CONFIG_LIBKLVANC
case AV_CODEC_ID_SMPTE_2038:
/* No specific setup required */
ret = 0;
break;
#endif
default:
av_log(avctx, AV_LOG_ERROR, "Unsupported data codec specified\n");
break;
}
return ret;
}
av_cold int ff_decklink_write_trailer(AVFormatContext *avctx)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
if (ctx->playback_started) {
BMDTimeValue actual;
ctx->dlo->StopScheduledPlayback(ctx->last_pts * ctx->bmd_tb_num,
&actual, ctx->bmd_tb_den);
ctx->dlo->DisableVideoOutput();
if (ctx->audio)
ctx->dlo->DisableAudioOutput();
}
ff_decklink_cleanup(avctx);
if (ctx->output_callback)
delete ctx->output_callback;
avdevice/decklink: remove pthread dependency Purpose: avdevice/decklink: Removed pthread dependency by replacing semaphore used in code appropriately. Doing so makes it easier to build ffmpeg using Visual C++ on Windows. This is a contination of Kyle Schwarz's "avdevice/decklink: Remove pthread dependency" patch that is available at https://patchwork.ffmpeg.org/patch/2654/ . This patch wasn't accepted, and as far as I can tell, there was no follow-up after it was rejected. Notes: Used Visual Studio 2015 (with update 3) for this. Comments: -- configure: Eliminated pthreads dependency for decklink_indev_deps and decklink_outdev_deps and replaced with threads dependency -- libavdevice/decklink_common.cpp / .h: a) Eliminated semaphore and replaced with a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). b) Removed include of pthread.h and semaphore.h and now using libavutil/thread.h instead. -- libavdevice/decklink_dec.cpp: Eliminated include of pthread.h and semaphore.h. -- libavdevice/decklink_enc.cpp: a) Eliminated include of pthread.h and semaphore.h. b) Replaced use of semaphore with the equivalent using a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). In theory, libavutil/thread.h and the associated code could have been modified instead to add cross-platform implementations of the sem_ functions, but an inspection of the ffmpeg source base indicates that there are only two cases in which semaphores are used (including this one that was replaced), so it was deemed to not be worth the effort. Signed-off-by: Marton Balint <cus@passwd.hu>
2017-04-15 08:41:46 +02:00
pthread_mutex_destroy(&ctx->mutex);
pthread_cond_destroy(&ctx->cond);
#if CONFIG_LIBKLVANC
klvanc_context_destroy(ctx->vanc_ctx);
#endif
ff_decklink_packet_queue_end(&ctx->vanc_queue);
ff_ccfifo_uninit(&ctx->cc_fifo);
av_freep(&cctx->ctx);
return 0;
}
#if CONFIG_LIBKLVANC
static void construct_cc(AVFormatContext *avctx, struct decklink_ctx *ctx,
AVPacket *pkt, struct klvanc_line_set_s *vanc_lines)
{
struct klvanc_packet_eia_708b_s *cdp;
uint16_t *cdp_words;
uint16_t len;
uint8_t cc_count;
size_t size;
int ret, i;
const uint8_t *data = av_packet_get_side_data(pkt, AV_PKT_DATA_A53_CC, &size);
if (!data)
return;
cc_count = size / 3;
ret = klvanc_create_eia708_cdp(&cdp);
if (ret)
return;
ret = klvanc_set_framerate_EIA_708B(cdp, ctx->bmd_tb_num, ctx->bmd_tb_den);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "Invalid framerate specified: %" PRId64 "/%" PRId64 "\n",
ctx->bmd_tb_num, ctx->bmd_tb_den);
klvanc_destroy_eia708_cdp(cdp);
return;
}
if (cc_count > KLVANC_MAX_CC_COUNT) {
av_log(avctx, AV_LOG_ERROR, "Illegal cc_count received: %d\n", cc_count);
cc_count = KLVANC_MAX_CC_COUNT;
}
/* CC data */
cdp->header.ccdata_present = 1;
cdp->header.caption_service_active = 1;
cdp->ccdata.cc_count = cc_count;
for (i = 0; i < cc_count; i++) {
if (data [3*i] & 0x04)
cdp->ccdata.cc[i].cc_valid = 1;
cdp->ccdata.cc[i].cc_type = data[3*i] & 0x03;
cdp->ccdata.cc[i].cc_data[0] = data[3*i+1];
cdp->ccdata.cc[i].cc_data[1] = data[3*i+2];
}
klvanc_finalize_EIA_708B(cdp, ctx->cdp_sequence_num++);
ret = klvanc_convert_EIA_708B_to_words(cdp, &cdp_words, &len);
klvanc_destroy_eia708_cdp(cdp);
if (ret != 0) {
av_log(avctx, AV_LOG_ERROR, "Failed converting 708 packet to words\n");
return;
}
ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, cdp_words, len, 11, 0);
free(cdp_words);
if (ret != 0) {
av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n");
return;
}
}
/* See SMPTE ST 2016-3:2009 */
static void construct_afd(AVFormatContext *avctx, struct decklink_ctx *ctx,
AVPacket *pkt, struct klvanc_line_set_s *vanc_lines,
AVStream *st)
{
struct klvanc_packet_afd_s *afd = NULL;
uint16_t *afd_words = NULL;
uint16_t len;
size_t size;
int f1_line = 12, f2_line = 0, ret;
const uint8_t *data = av_packet_get_side_data(pkt, AV_PKT_DATA_AFD, &size);
if (!data || size == 0)
return;
ret = klvanc_create_AFD(&afd);
if (ret)
return;
ret = klvanc_set_AFD_val(afd, data[0]);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "Invalid AFD value specified: %d\n",
data[0]);
klvanc_destroy_AFD(afd);
return;
}
/* Compute the AR flag based on the DAR (see ST 2016-1:2009 Sec 9.1). Note, we treat
anything below 1.4 as 4:3 (as opposed to the standard 1.33), because there are lots
of streams in the field that aren't *exactly* 4:3 but a tiny bit larger after doing
the math... */
if (av_cmp_q((AVRational) {st->codecpar->width * st->codecpar->sample_aspect_ratio.num,
st->codecpar->height * st->codecpar->sample_aspect_ratio.den}, (AVRational) {14, 10}) == 1)
afd->aspectRatio = ASPECT_16x9;
else
afd->aspectRatio = ASPECT_4x3;
ret = klvanc_convert_AFD_to_words(afd, &afd_words, &len);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "Failed converting AFD packet to words\n");
goto out;
}
ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, afd_words, len, f1_line, 0);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n");
goto out;
}
/* For interlaced video, insert into both fields. Switching lines for field 2
derived from SMPTE RP 168:2009, Sec 6, Table 2. */
switch (ctx->bmd_mode) {
case bmdModeNTSC:
case bmdModeNTSC2398:
f2_line = 273 - 10 + f1_line;
break;
case bmdModePAL:
f2_line = 319 - 6 + f1_line;
break;
case bmdModeHD1080i50:
case bmdModeHD1080i5994:
case bmdModeHD1080i6000:
f2_line = 569 - 7 + f1_line;
break;
default:
f2_line = 0;
break;
}
if (f2_line > 0) {
ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, afd_words, len, f2_line, 0);
if (ret) {
av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n");
goto out;
}
}
out:
if (afd)
klvanc_destroy_AFD(afd);
if (afd_words)
free(afd_words);
}
/* Parse any EIA-608 subtitles sitting on the queue, and write packet side data
that will later be handled by construct_cc... */
static void parse_608subs(AVFormatContext *avctx, struct decklink_ctx *ctx, AVPacket *pkt)
{
size_t cc_size = ff_ccfifo_getoutputsize(&ctx->cc_fifo);
uint8_t *cc_data;
if (!ff_ccfifo_ccdetected(&ctx->cc_fifo))
return;
cc_data = av_packet_new_side_data(pkt, AV_PKT_DATA_A53_CC, cc_size);
if (cc_data)
ff_ccfifo_injectbytes(&ctx->cc_fifo, cc_data, cc_size);
}
static int decklink_construct_vanc(AVFormatContext *avctx, struct decklink_ctx *ctx,
AVPacket *pkt, decklink_frame *frame,
AVStream *st)
{
struct klvanc_line_set_s vanc_lines = { 0 };
int ret = 0, i;
if (!ctx->supports_vanc)
return 0;
parse_608subs(avctx, ctx, pkt);
construct_cc(avctx, ctx, pkt, &vanc_lines);
construct_afd(avctx, ctx, pkt, &vanc_lines, st);
/* See if there any pending data packets to process */
while (ff_decklink_packet_queue_size(&ctx->vanc_queue) > 0) {
AVStream *vanc_st;
AVPacket vanc_pkt;
int64_t pts;
pts = ff_decklink_packet_queue_peekpts(&ctx->vanc_queue);
if (pts > ctx->last_pts) {
/* We haven't gotten to the video frame we are supposed to inject
the oldest VANC packet into yet, so leave it on the queue... */
break;
}
ret = ff_decklink_packet_queue_get(&ctx->vanc_queue, &vanc_pkt, 1);
if (vanc_pkt.pts + 1 < ctx->last_pts) {
av_log(avctx, AV_LOG_WARNING, "VANC packet too old, throwing away\n");
av_packet_unref(&vanc_pkt);
continue;
}
vanc_st = avctx->streams[vanc_pkt.stream_index];
if (vanc_st->codecpar->codec_id == AV_CODEC_ID_SMPTE_2038) {
struct klvanc_smpte2038_anc_data_packet_s *pkt_2038 = NULL;
klvanc_smpte2038_parse_pes_payload(vanc_pkt.data, vanc_pkt.size, &pkt_2038);
if (pkt_2038 == NULL) {
av_log(avctx, AV_LOG_ERROR, "failed to decode SMPTE 2038 PES packet");
av_packet_unref(&vanc_pkt);
continue;
}
for (int i = 0; i < pkt_2038->lineCount; i++) {
struct klvanc_smpte2038_anc_data_line_s *l = &pkt_2038->lines[i];
uint16_t *vancWords = NULL;
uint16_t vancWordCount;
if (klvanc_smpte2038_convert_line_to_words(l, &vancWords,
&vancWordCount) < 0)
break;
ret = klvanc_line_insert(ctx->vanc_ctx, &vanc_lines, vancWords,
vancWordCount, l->line_number, 0);
free(vancWords);
if (ret != 0) {
av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n");
break;
}
}
klvanc_smpte2038_anc_data_packet_free(pkt_2038);
}
av_packet_unref(&vanc_pkt);
}
IDeckLinkVideoFrameAncillary *vanc;
int result = ctx->dlo->CreateAncillaryData(bmdFormat10BitYUV, &vanc);
if (result != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Failed to create vanc\n");
ret = AVERROR(EIO);
goto done;
}
/* Now that we've got all the VANC lines in a nice orderly manner, generate the
final VANC sections for the Decklink output */
for (i = 0; i < vanc_lines.num_lines; i++) {
struct klvanc_line_s *line = vanc_lines.lines[i];
int real_line;
void *buf;
if (!line)
break;
/* FIXME: include hack for certain Decklink cards which mis-represent
line numbers for pSF frames */
real_line = line->line_number;
result = vanc->GetBufferForVerticalBlankingLine(real_line, &buf);
if (result != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Failed to get VANC line %d: %d", real_line, result);
continue;
}
/* Generate the full line taking into account all VANC packets on that line */
result = klvanc_generate_vanc_line_v210(ctx->vanc_ctx, line, (uint8_t *) buf,
ctx->bmd_width);
if (result) {
av_log(avctx, AV_LOG_ERROR, "Failed to generate VANC line\n");
continue;
}
}
result = frame->SetAncillaryData(vanc);
vanc->Release();
if (result != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Failed to set vanc: %d", result);
ret = AVERROR(EIO);
}
done:
for (i = 0; i < vanc_lines.num_lines; i++)
klvanc_line_free(vanc_lines.lines[i]);
return ret;
}
#endif
static int decklink_write_video_packet(AVFormatContext *avctx, AVPacket *pkt)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
AVStream *st = avctx->streams[pkt->stream_index];
AVFrame *avframe = NULL, *tmp = (AVFrame *)pkt->data;
AVPacket *avpacket = NULL;
decklink_frame *frame;
uint32_t buffered;
HRESULT hr;
ctx->last_pts = FFMAX(ctx->last_pts, pkt->pts);
if (st->codecpar->codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) {
if (tmp->format != AV_PIX_FMT_UYVY422 ||
tmp->width != ctx->bmd_width ||
tmp->height != ctx->bmd_height) {
av_log(avctx, AV_LOG_ERROR, "Got a frame with invalid pixel format or dimension.\n");
return AVERROR(EINVAL);
}
avframe = av_frame_clone(tmp);
if (!avframe) {
av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n");
return AVERROR(EIO);
}
frame = new decklink_frame(ctx, avframe, st->codecpar->codec_id, avframe->height, avframe->width);
} else {
avpacket = av_packet_clone(pkt);
if (!avpacket) {
av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n");
return AVERROR(EIO);
}
frame = new decklink_frame(ctx, avpacket, st->codecpar->codec_id, ctx->bmd_height, ctx->bmd_width);
#if CONFIG_LIBKLVANC
if (decklink_construct_vanc(avctx, ctx, pkt, frame, st))
av_log(avctx, AV_LOG_ERROR, "Failed to construct VANC\n");
#endif
}
if (!frame) {
av_log(avctx, AV_LOG_ERROR, "Could not create new frame.\n");
av_frame_free(&avframe);
av_packet_free(&avpacket);
return AVERROR(EIO);
}
/* Always keep at most one second of frames buffered. */
avdevice/decklink: remove pthread dependency Purpose: avdevice/decklink: Removed pthread dependency by replacing semaphore used in code appropriately. Doing so makes it easier to build ffmpeg using Visual C++ on Windows. This is a contination of Kyle Schwarz's "avdevice/decklink: Remove pthread dependency" patch that is available at https://patchwork.ffmpeg.org/patch/2654/ . This patch wasn't accepted, and as far as I can tell, there was no follow-up after it was rejected. Notes: Used Visual Studio 2015 (with update 3) for this. Comments: -- configure: Eliminated pthreads dependency for decklink_indev_deps and decklink_outdev_deps and replaced with threads dependency -- libavdevice/decklink_common.cpp / .h: a) Eliminated semaphore and replaced with a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). b) Removed include of pthread.h and semaphore.h and now using libavutil/thread.h instead. -- libavdevice/decklink_dec.cpp: Eliminated include of pthread.h and semaphore.h. -- libavdevice/decklink_enc.cpp: a) Eliminated include of pthread.h and semaphore.h. b) Replaced use of semaphore with the equivalent using a combination of a mutex, condition variable, and a counter (frames_buffer_available_spots). In theory, libavutil/thread.h and the associated code could have been modified instead to add cross-platform implementations of the sem_ functions, but an inspection of the ffmpeg source base indicates that there are only two cases in which semaphores are used (including this one that was replaced), so it was deemed to not be worth the effort. Signed-off-by: Marton Balint <cus@passwd.hu>
2017-04-15 08:41:46 +02:00
pthread_mutex_lock(&ctx->mutex);
while (ctx->frames_buffer_available_spots == 0) {
pthread_cond_wait(&ctx->cond, &ctx->mutex);
}
ctx->frames_buffer_available_spots--;
pthread_mutex_unlock(&ctx->mutex);
if (ctx->first_pts == AV_NOPTS_VALUE)
ctx->first_pts = pkt->pts;
/* Schedule frame for playback. */
hr = ctx->dlo->ScheduleVideoFrame((class IDeckLinkVideoFrame *) frame,
pkt->pts * ctx->bmd_tb_num,
ctx->bmd_tb_num, ctx->bmd_tb_den);
/* Pass ownership to DeckLink, or release on failure */
frame->Release();
if (hr != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not schedule video frame."
" error %08x.\n", (uint32_t) hr);
return AVERROR(EIO);
}
ctx->dlo->GetBufferedVideoFrameCount(&buffered);
av_log(avctx, AV_LOG_DEBUG, "Buffered video frames: %d.\n", (int) buffered);
if (pkt->pts > 2 && buffered <= 2)
av_log(avctx, AV_LOG_WARNING, "There are not enough buffered video frames."
" Video may misbehave!\n");
/* Preroll video frames. */
if (!ctx->playback_started && pkt->pts > (ctx->first_pts + ctx->frames_preroll)) {
av_log(avctx, AV_LOG_DEBUG, "Ending audio preroll.\n");
if (ctx->audio && ctx->dlo->EndAudioPreroll() != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not end audio preroll!\n");
return AVERROR(EIO);
}
av_log(avctx, AV_LOG_DEBUG, "Starting scheduled playback.\n");
if (ctx->dlo->StartScheduledPlayback(ctx->first_pts * ctx->bmd_tb_num, ctx->bmd_tb_den, 1.0) != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not start scheduled playback!\n");
return AVERROR(EIO);
}
ctx->playback_started = 1;
}
return 0;
}
static int decklink_write_audio_packet(AVFormatContext *avctx, AVPacket *pkt)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
AVStream *st = avctx->streams[pkt->stream_index];
int sample_count;
uint32_t buffered;
uint8_t *outbuf = NULL;
int ret = 0;
ctx->dlo->GetBufferedAudioSampleFrameCount(&buffered);
if (pkt->pts > 1 && !buffered)
av_log(avctx, AV_LOG_WARNING, "There's no buffered audio."
" Audio will misbehave!\n");
if (st->codecpar->codec_id == AV_CODEC_ID_AC3) {
/* Encapsulate AC3 syncframe into SMPTE 337 packet */
int outbuf_size;
ret = create_s337_payload(pkt, &outbuf, &outbuf_size);
if (ret < 0)
return ret;
sample_count = outbuf_size / 4;
} else {
sample_count = pkt->size / (ctx->channels << 1);
outbuf = pkt->data;
}
if (ctx->dlo->ScheduleAudioSamples(outbuf, sample_count, pkt->pts,
bmdAudioSampleRate48kHz, NULL) != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not schedule audio samples.\n");
ret = AVERROR(EIO);
}
if (st->codecpar->codec_id == AV_CODEC_ID_AC3)
av_freep(&outbuf);
return ret;
}
static int decklink_write_subtitle_packet(AVFormatContext *avctx, AVPacket *pkt)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
ff_ccfifo_extractbytes(&ctx->cc_fifo, pkt->data, pkt->size);
return 0;
}
static int decklink_write_data_packet(AVFormatContext *avctx, AVPacket *pkt)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx;
if (ff_decklink_packet_queue_put(&ctx->vanc_queue, pkt) < 0) {
av_log(avctx, AV_LOG_WARNING, "Failed to queue DATA packet\n");
}
return 0;
}
extern "C" {
av_cold int ff_decklink_write_header(AVFormatContext *avctx)
{
struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data;
struct decklink_ctx *ctx;
unsigned int n;
int ret;
ctx = (struct decklink_ctx *) av_mallocz(sizeof(struct decklink_ctx));
if (!ctx)
return AVERROR(ENOMEM);
ctx->list_devices = cctx->list_devices;
ctx->list_formats = cctx->list_formats;
ctx->preroll = cctx->preroll;
ctx->duplex_mode = cctx->duplex_mode;
ctx->first_pts = AV_NOPTS_VALUE;
if (cctx->link > 0 && (unsigned int)cctx->link < FF_ARRAY_ELEMS(decklink_link_conf_map))
ctx->link = decklink_link_conf_map[cctx->link];
cctx->ctx = ctx;
#if CONFIG_LIBKLVANC
if (klvanc_context_create(&ctx->vanc_ctx) < 0) {
av_log(avctx, AV_LOG_ERROR, "Cannot create VANC library context\n");
return AVERROR(ENOMEM);
}
ctx->supports_vanc = 1;
#endif
/* List available devices and exit. */
if (ctx->list_devices) {
ff_decklink_list_devices_legacy(avctx, 0, 1);
return AVERROR_EXIT;
}
ret = ff_decklink_init_device(avctx, avctx->url);
if (ret < 0)
return ret;
/* Get output device. */
if (ctx->dl->QueryInterface(IID_IDeckLinkOutput, (void **) &ctx->dlo) != S_OK) {
av_log(avctx, AV_LOG_ERROR, "Could not open output device from '%s'\n",
avctx->url);
ret = AVERROR(EIO);
goto error;
}
/* List supported formats. */
if (ctx->list_formats) {
ff_decklink_list_formats(avctx);
ret = AVERROR_EXIT;
goto error;
}
/* Setup streams. */
ret = AVERROR(EIO);
for (n = 0; n < avctx->nb_streams; n++) {
AVStream *st = avctx->streams[n];
AVCodecParameters *c = st->codecpar;
if (c->codec_type == AVMEDIA_TYPE_AUDIO) {
if (decklink_setup_audio(avctx, st))
goto error;
} else if (c->codec_type == AVMEDIA_TYPE_VIDEO) {
if (decklink_setup_video(avctx, st))
goto error;
} else if (c->codec_type == AVMEDIA_TYPE_DATA) {
if (decklink_setup_data(avctx, st))
goto error;
} else if (c->codec_type == AVMEDIA_TYPE_SUBTITLE) {
if (decklink_setup_subtitle(avctx, st))
goto error;
} else {
av_log(avctx, AV_LOG_ERROR, "Unsupported stream type.\n");
goto error;
}
}
/* Reconfigure the data/subtitle stream clocks to match the video */
for (n = 0; n < avctx->nb_streams; n++) {
AVStream *st = avctx->streams[n];
AVCodecParameters *c = st->codecpar;
if(c->codec_type == AVMEDIA_TYPE_DATA ||
c->codec_type == AVMEDIA_TYPE_SUBTITLE)
avpriv_set_pts_info(st, 64, ctx->bmd_tb_num, ctx->bmd_tb_den);
}
ff_decklink_packet_queue_init(avctx, &ctx->vanc_queue, cctx->vanc_queue_size);
ret = ff_ccfifo_init(&ctx->cc_fifo, av_make_q(ctx->bmd_tb_den, ctx->bmd_tb_num), avctx);
if (ret < 0) {
av_log(ctx, AV_LOG_ERROR, "Failure to setup CC FIFO queue\n");
goto error;
}
return 0;
error:
ff_decklink_cleanup(avctx);
return ret;
}
int ff_decklink_write_packet(AVFormatContext *avctx, AVPacket *pkt)
{
AVStream *st = avctx->streams[pkt->stream_index];
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
return decklink_write_video_packet(avctx, pkt);
else if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO)
return decklink_write_audio_packet(avctx, pkt);
else if (st->codecpar->codec_type == AVMEDIA_TYPE_DATA)
return decklink_write_data_packet(avctx, pkt);
else if (st->codecpar->codec_type == AVMEDIA_TYPE_SUBTITLE)
return decklink_write_subtitle_packet(avctx, pkt);
return AVERROR(EIO);
}
int ff_decklink_list_output_devices(AVFormatContext *avctx, struct AVDeviceInfoList *device_list)
{
return ff_decklink_list_devices(avctx, device_list, 0, 1);
}
} /* extern "C" */