/* * Blackmagic DeckLink output * Copyright (c) 2013-2014 Ramiro Polla * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include extern "C" { #include "libavformat/avformat.h" #include "libavformat/internal.h" #include "libavutil/imgutils.h" } #include "decklink_common.h" #include "decklink_enc.h" /* DeckLink callback class declaration */ class decklink_frame : public IDeckLinkVideoFrame { public: decklink_frame(struct decklink_ctx *ctx, AVFrame *avframe, long width, long height, void *buffer) : _ctx(ctx), _avframe(avframe), _width(width), _height(height), _buffer(buffer), _refs(0) { } virtual long STDMETHODCALLTYPE GetWidth (void) { return _width; } virtual long STDMETHODCALLTYPE GetHeight (void) { return _height; } virtual long STDMETHODCALLTYPE GetRowBytes (void) { return _width<<1; } virtual BMDPixelFormat STDMETHODCALLTYPE GetPixelFormat(void) { return bmdFormat8BitYUV; } virtual BMDFrameFlags STDMETHODCALLTYPE GetFlags (void) { return bmdVideoOutputFlagDefault; } virtual HRESULT STDMETHODCALLTYPE GetBytes (void **buffer) { *buffer = _buffer; return S_OK; } virtual HRESULT STDMETHODCALLTYPE GetTimecode (BMDTimecodeFormat format, IDeckLinkTimecode **timecode) { return S_FALSE; } virtual HRESULT STDMETHODCALLTYPE GetAncillaryData(IDeckLinkVideoFrameAncillary **ancillary) { return S_FALSE; } virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void) { return ++_refs; } virtual ULONG STDMETHODCALLTYPE Release(void) { if (!--_refs) delete this; return _refs; } struct decklink_ctx *_ctx; AVFrame *_avframe; private: long _width; long _height; void *_buffer; int _refs; }; class decklink_output_callback : public IDeckLinkVideoOutputCallback { public: virtual HRESULT STDMETHODCALLTYPE ScheduledFrameCompleted(IDeckLinkVideoFrame *_frame, BMDOutputFrameCompletionResult result) { decklink_frame *frame = static_cast(_frame); struct decklink_ctx *ctx = frame->_ctx; AVFrame *avframe = frame->_avframe; av_frame_free(&avframe); sem_post(&ctx->semaphore); return S_OK; } virtual HRESULT STDMETHODCALLTYPE ScheduledPlaybackHasStopped(void) { return S_OK; } virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void) { return 1; } virtual ULONG STDMETHODCALLTYPE Release(void) { return 1; } }; static int decklink_setup_video(AVFormatContext *avctx, AVStream *st) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; AVCodecContext *c = st->codec; if (ctx->video) { av_log(avctx, AV_LOG_ERROR, "Only one video stream is supported!\n"); return -1; } if (c->pix_fmt != AV_PIX_FMT_UYVY422) { av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format!" " Only AV_PIX_FMT_UYVY422 is supported.\n"); return -1; } if (ff_decklink_set_format(avctx, c->width, c->height, c->time_base.num, c->time_base.den)) { av_log(avctx, AV_LOG_ERROR, "Unsupported video size or framerate!" " Check available formats with -list_formats 1.\n"); return -1; } if (ctx->dlo->EnableVideoOutput(ctx->bmd_mode, bmdVideoOutputFlagDefault) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not enable video output!\n"); return -1; } /* Set callback. */ ctx->output_callback = new decklink_output_callback(); ctx->dlo->SetScheduledFrameCompletionCallback(ctx->output_callback); /* Start video semaphore. */ ctx->frames_preroll = c->time_base.den * ctx->preroll; if (c->time_base.den > 1000) ctx->frames_preroll /= 1000; /* Buffer twice as many frames as the preroll. */ ctx->frames_buffer = ctx->frames_preroll * 2; ctx->frames_buffer = FFMIN(ctx->frames_buffer, 60); sem_init(&ctx->semaphore, 0, ctx->frames_buffer); /* The device expects the framerate to be fixed. */ avpriv_set_pts_info(st, 64, c->time_base.num, c->time_base.den); ctx->video = 1; return 0; } static int decklink_setup_audio(AVFormatContext *avctx, AVStream *st) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; AVCodecContext *c = st->codec; if (ctx->audio) { av_log(avctx, AV_LOG_ERROR, "Only one audio stream is supported!\n"); return -1; } if (c->sample_rate != 48000) { av_log(avctx, AV_LOG_ERROR, "Unsupported sample rate!" " Only 48kHz is supported.\n"); return -1; } if (c->channels != 2 && c->channels != 8) { av_log(avctx, AV_LOG_ERROR, "Unsupported number of channels!" " Only stereo and 7.1 are supported.\n"); return -1; } if (ctx->dlo->EnableAudioOutput(bmdAudioSampleRate48kHz, bmdAudioSampleType16bitInteger, c->channels, bmdAudioOutputStreamTimestamped) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not enable audio output!\n"); return -1; } if (ctx->dlo->BeginAudioPreroll() != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not begin audio preroll!\n"); return -1; } /* The device expects the sample rate to be fixed. */ avpriv_set_pts_info(st, 64, 1, c->sample_rate); ctx->channels = c->channels; ctx->audio = 1; return 0; } av_cold int ff_decklink_write_trailer(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; if (ctx->playback_started) { BMDTimeValue actual; ctx->dlo->StopScheduledPlayback(ctx->last_pts * ctx->bmd_tb_num, &actual, ctx->bmd_tb_den); ctx->dlo->DisableVideoOutput(); if (ctx->audio) ctx->dlo->DisableAudioOutput(); } if (ctx->dlo) ctx->dlo->Release(); if (ctx->dl) ctx->dl->Release(); if (ctx->output_callback) delete ctx->output_callback; sem_destroy(&ctx->semaphore); av_freep(&cctx->ctx); return 0; } static int decklink_write_video_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; AVPicture *avpicture = (AVPicture *) pkt->data; AVFrame *avframe, *tmp; decklink_frame *frame; buffercount_type buffered; HRESULT hr; /* HACK while av_uncoded_frame() isn't implemented */ int ret; tmp = av_frame_alloc(); if (!tmp) return AVERROR(ENOMEM); tmp->format = AV_PIX_FMT_UYVY422; tmp->width = ctx->bmd_width; tmp->height = ctx->bmd_height; ret = av_frame_get_buffer(tmp, 32); if (ret < 0) { av_frame_free(&tmp); return ret; } av_image_copy(tmp->data, tmp->linesize, (const uint8_t **) avpicture->data, avpicture->linesize, (AVPixelFormat) tmp->format, tmp->width, tmp->height); avframe = av_frame_clone(tmp); av_frame_free(&tmp); if (!avframe) { av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n"); return AVERROR(EIO); } /* end HACK */ frame = new decklink_frame(ctx, avframe, ctx->bmd_width, ctx->bmd_height, (void *) avframe->data[0]); if (!frame) { av_log(avctx, AV_LOG_ERROR, "Could not create new frame.\n"); return AVERROR(EIO); } /* Always keep at most one second of frames buffered. */ sem_wait(&ctx->semaphore); /* Schedule frame for playback. */ hr = ctx->dlo->ScheduleVideoFrame((struct IDeckLinkVideoFrame *) frame, pkt->pts * ctx->bmd_tb_num, ctx->bmd_tb_num, ctx->bmd_tb_den); if (hr != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not schedule video frame." " error %08x.\n", (uint32_t) hr); frame->Release(); return AVERROR(EIO); } ctx->dlo->GetBufferedVideoFrameCount(&buffered); av_log(avctx, AV_LOG_DEBUG, "Buffered video frames: %d.\n", (int) buffered); if (pkt->pts > 2 && buffered <= 2) av_log(avctx, AV_LOG_WARNING, "There are not enough buffered video frames." " Video may misbehave!\n"); /* Preroll video frames. */ if (!ctx->playback_started && pkt->pts > ctx->frames_preroll) { av_log(avctx, AV_LOG_DEBUG, "Ending audio preroll.\n"); if (ctx->audio && ctx->dlo->EndAudioPreroll() != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not end audio preroll!\n"); return AVERROR(EIO); } av_log(avctx, AV_LOG_DEBUG, "Starting scheduled playback.\n"); if (ctx->dlo->StartScheduledPlayback(0, ctx->bmd_tb_den, 1.0) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not start scheduled playback!\n"); return AVERROR(EIO); } ctx->playback_started = 1; } return 0; } static int decklink_write_audio_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; int sample_count = pkt->size / (ctx->channels << 1); buffercount_type buffered; ctx->dlo->GetBufferedAudioSampleFrameCount(&buffered); if (pkt->pts > 1 && !buffered) av_log(avctx, AV_LOG_WARNING, "There's no buffered audio." " Audio will misbehave!\n"); if (ctx->dlo->ScheduleAudioSamples(pkt->data, sample_count, pkt->pts, bmdAudioSampleRate48kHz, NULL) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not schedule audio samples.\n"); return AVERROR(EIO); } return 0; } extern "C" { av_cold int ff_decklink_write_header(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx; IDeckLinkDisplayModeIterator *itermode; IDeckLinkIterator *iter; IDeckLink *dl = NULL; unsigned int n; ctx = (struct decklink_ctx *) av_mallocz(sizeof(struct decklink_ctx)); if (!ctx) return AVERROR(ENOMEM); ctx->list_devices = cctx->list_devices; ctx->list_formats = cctx->list_formats; ctx->preroll = cctx->preroll; cctx->ctx = ctx; iter = CreateDeckLinkIteratorInstance(); if (!iter) { av_log(avctx, AV_LOG_ERROR, "Could not create DeckLink iterator\n"); return AVERROR(EIO); } /* List available devices. */ if (ctx->list_devices) { ff_decklink_list_devices(avctx); return AVERROR_EXIT; } /* Open device. */ while (iter->Next(&dl) == S_OK) { const char *displayName; ff_decklink_get_display_name(dl, &displayName); if (!strcmp(avctx->filename, displayName)) { av_free((void *) displayName); ctx->dl = dl; break; } av_free((void *) displayName); dl->Release(); } iter->Release(); if (!ctx->dl) { av_log(avctx, AV_LOG_ERROR, "Could not open '%s'\n", avctx->filename); return AVERROR(EIO); } /* Get output device. */ if (ctx->dl->QueryInterface(IID_IDeckLinkOutput, (void **) &ctx->dlo) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not open output device from '%s'\n", avctx->filename); ctx->dl->Release(); return AVERROR(EIO); } /* List supported formats. */ if (ctx->list_formats) { ff_decklink_list_formats(avctx); ctx->dlo->Release(); ctx->dl->Release(); return AVERROR_EXIT; } if (ctx->dlo->GetDisplayModeIterator(&itermode) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not get Display Mode Iterator\n"); ctx->dl->Release(); return AVERROR(EIO); } /* Setup streams. */ for (n = 0; n < avctx->nb_streams; n++) { AVStream *st = avctx->streams[n]; AVCodecContext *c = st->codec; if (c->codec_type == AVMEDIA_TYPE_AUDIO) { if (decklink_setup_audio(avctx, st)) goto error; } else if (c->codec_type == AVMEDIA_TYPE_VIDEO) { if (decklink_setup_video(avctx, st)) goto error; } else { av_log(avctx, AV_LOG_ERROR, "Unsupported stream type.\n"); goto error; } } itermode->Release(); return 0; error: ctx->dlo->Release(); ctx->dl->Release(); return AVERROR(EIO); } int ff_decklink_write_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; AVStream *st = avctx->streams[pkt->stream_index]; ctx->last_pts = FFMAX(ctx->last_pts, pkt->pts); if (st->codec->codec_type == AVMEDIA_TYPE_VIDEO) return decklink_write_video_packet(avctx, pkt); else if (st->codec->codec_type == AVMEDIA_TYPE_AUDIO) return decklink_write_audio_packet(avctx, pkt); return AVERROR(EIO); } } /* extern "C" */