34 #include <DeckLinkAPI.h> 57 #define MAX_WIDTH_VANC 1920 76 {bmdModeNTSC, 11, 19, 274, 282},
77 {bmdModeNTSC2398, 11, 19, 274, 282},
78 {bmdModePAL, 7, 22, 320, 335},
79 {bmdModeNTSCp, 11, -1, -1, 39},
80 {bmdModePALp, 7, -1, -1, 45},
84 {bmdModeHD1080p2398, 8, -1, -1, 42},
85 {bmdModeHD1080p24, 8, -1, -1, 42},
86 {bmdModeHD1080p25, 8, -1, -1, 42},
87 {bmdModeHD1080p2997, 8, -1, -1, 42},
88 {bmdModeHD1080p30, 8, -1, -1, 42},
89 {bmdModeHD1080i50, 8, 20, 570, 585},
90 {bmdModeHD1080i5994, 8, 20, 570, 585},
91 {bmdModeHD1080i6000, 8, 20, 570, 585},
92 {bmdModeHD1080p50, 8, -1, -1, 42},
93 {bmdModeHD1080p5994, 8, -1, -1, 42},
94 {bmdModeHD1080p6000, 8, -1, -1, 42},
98 {bmdModeHD720p50, 8, -1, -1, 26},
99 {bmdModeHD720p5994, 8, -1, -1, 26},
100 {bmdModeHD720p60, 8, -1, -1, 26},
103 {bmdModeUnknown, 0, -1, -1, -1}
113 virtual HRESULT STDMETHODCALLTYPE
AllocateBuffer(
unsigned int bufferSize,
void* *allocatedBuffer)
117 return E_OUTOFMEMORY;
118 *allocatedBuffer = buf;
126 virtual HRESULT STDMETHODCALLTYPE
Commit() {
return S_OK; }
127 virtual HRESULT STDMETHODCALLTYPE
Decommit() {
return S_OK; }
130 virtual HRESULT STDMETHODCALLTYPE
QueryInterface(REFIID iid, LPVOID *ppv) {
return E_NOINTERFACE; }
131 virtual ULONG STDMETHODCALLTYPE
AddRef(
void) {
return ++_refs; }
147 IUnknown *obj = (
class IUnknown *)opaque;
156 if (mode == vanc_line_numbers[i].mode)
165 for (i = 0; i <
len; i++)
171 uint16_t vanc_sum = 0;
172 for (i = 3; i < len - 1; i++) {
176 if ((!!p ^ !!(v & 0x100)) || (np != 1 && np != 2)) {
183 vanc_sum |= ((~vanc_sum & 0x100) << 1);
184 if (checksum != vanc_sum) {
195 for (i = 0; i < width / 3; i++) {
196 *dst++ = (src[1] >> 2) + ((src[2] & 15) << 6);
197 *dst++ = src[4] + ((src[5] & 3) << 8);
198 *dst++ = (src[6] >> 4) + ((src[7] & 63) << 4);
206 for (i = 0; i < width * 2 / 3; i++) {
207 *dst++ = src[0] + ((src[1] & 3) << 8);
208 *dst++ = (src[1] >> 2) + ((src[2] & 15) << 6);
209 *dst++ = (src[2] >> 4) + ((src[3] & 63) << 4);
217 if (line >= 7 && line <= 22)
219 if (line >= 320 && line <= 335)
235 vbi_bit_slicer slicer;
237 vbi_bit_slicer_init(&slicer, 720, 13500000, 6937500, 6937500, 0x00aaaae4, 0xffff, 18, 6, 42 * 8, VBI_MODULATION_NRZ_MSB, fmt);
239 if (vbi_bit_slice(&slicer,
src, tgt + 4) == FALSE)
256 *py++ = (
src[1] >> 4) + ((
src[2] & 15) << 4);
257 *py++ = (
src[4] >> 2) + ((
src[5] & 3 ) << 6);
258 *py++ = (
src[6] >> 6) + ((
src[7] & 63) << 2);
261 return teletext_data_unit_from_vbi_data(
line, y, tgt, VBI_PIXFMT_YUV420);
269 if (py[0] != 0x255 || py[1] != 0x255 || py[2] != 0x227)
277 for (i = 0; i < 42; i++)
286 if (line >= 6 && line <= 22)
288 if (line >= 318 && line <= 335)
289 shift = line - 318 + 17;
290 return shift >= 0 && ((1ULL <<
shift) & mask);
296 if (py[0] == 0x151 && py[1] == 0x115 && py[3] == 0x102) {
297 uint16_t *descriptors = py + 4;
300 for (i = 0; i < 5 && py < pend - 45; i++, py += 45) {
301 int line = (descriptors[
i] & 31) + (!(descriptors[i] & 128)) * 313;
312 uint16_t did = py[0];
313 uint16_t sdid = py[1];
314 uint16_t
dc = py[2] & 255;
316 pend =
FFMIN(pend, py + dc);
317 if (did == 0x143 && sdid == 0x102) {
319 }
else if (allow_multipacket && did == 0x143 && sdid == 0x203) {
321 while (py < pend - 3) {
323 py += 4 + (py[2] & 255);
332 size_t i,
len = (buf[5] & 0xff) + 6 + 1;
336 uint16_t *cdp = &buf[6];
337 if (cdp[0] != 0x96 || cdp[1] != 0x69) {
350 for (i = 0; i < len - 1; i++)
352 cdp_sum = cdp_sum ? 256 - cdp_sum : 0;
353 if (cdp[len - 1] != cdp_sum) {
359 if (!(rate & 0x0f)) {
369 if (!(cdp[4] & 0x43)) {
374 hdr = (cdp[5] << 8) | cdp[6];
375 if (cdp[7] != 0x72) {
381 if (!(cc_count & 0xe0)) {
387 if ((len - 13) < cc_count * 3) {
392 if (cdp[len - 4] != 0x74) {
397 ftr = (cdp[len - 3] << 8) | cdp[len - 2];
409 for (
size_t i = 0; i < cc_count; i++) {
410 cc[3*i + 0] = cdp[9 + 3*i+0] ;
411 cc[3*i + 1] = cdp[9 + 3*i+1];
412 cc[3*i + 2] = cdp[9 + 3*i+2];
423 uint16_t *max_buf = buf + width;
425 while (buf < max_buf - 6) {
427 uint16_t did = buf[3] & 0xFF;
428 uint16_t sdid = buf[4] & 0xFF;
430 if (buf[0] != 0 || buf[1] != 0x3ff || buf[2] != 0x3ff) {
434 len = (buf[5] & 0xff) + 6 + 1;
435 if (len > max_buf - buf) {
441 if (did == 0x43 && (sdid == 0x02 || sdid == 0x03) && cctx->
teletext_lines &&
442 width == 1920 && tgt_size >= 1920) {
448 }
else if (did == 0x61 && sdid == 0x01) {
449 unsigned int data_len;
456 data =
vanc_to_cc(avctx, buf, width, data_len);
508 unsigned long long size;
591 const uint8_t KLV_IN_VANC_SDID = 0x04;
595 uint16_t sequence_counter;
596 std::vector<uint8_t>
data;
599 size_t total_size = 0;
600 std::vector<std::vector<KLVPacket>> klv_packets(256);
602 IDeckLinkVideoFrameAncillaryPackets *packets =
nullptr;
603 if (videoFrame->QueryInterface(IID_IDeckLinkVideoFrameAncillaryPackets, (
void**)&packets) != S_OK)
606 IDeckLinkAncillaryPacketIterator *
it =
nullptr;
607 if (packets->GetPacketIterator(&it) != S_OK) {
612 IDeckLinkAncillaryPacket *packet =
nullptr;
613 while (it->Next(&packet) == S_OK) {
617 if (packet->GetDID() == KLV_DID && packet->GetSDID() == KLV_IN_VANC_SDID) {
618 av_log(avctx,
AV_LOG_DEBUG,
"Found KLV VANC packet on line: %d\n", packet->GetLineNumber());
620 if (packet->GetBytes(bmdAncillaryPacketFormatUInt8, (
const void**) &data, &size) == S_OK) {
624 uint16_t psc = data[1] << 8 | data[2];
628 auto&
list = klv_packets[mid];
629 uint16_t expected_psc =
list.size() + 1;
631 if (psc == expected_psc) {
632 uint32_t data_len = size - 3;
633 total_size += data_len;
636 packet.data.resize(data_len);
637 memcpy(packet.data.data(), data + 3, data_len);
639 list.push_back(std::move(packet));
644 for (
auto& klv :
list)
645 total_size -= klv.data.size();
660 if (total_size > 0) {
661 std::vector<uint8_t> klv;
662 klv.reserve(total_size);
664 for (
size_t i = 0;
i < klv_packets.size(); ++
i) {
665 auto&
list = klv_packets[
i];
672 for (
auto& packet :
list)
673 klv.insert(klv.end(), packet.data.begin(), packet.data.end());
682 klv_packet.
data = klv.data();
683 klv_packet.
size = klv.size();
697 virtual HRESULT STDMETHODCALLTYPE
QueryInterface(REFIID iid, LPVOID *ppv) {
return E_NOINTERFACE; }
698 virtual ULONG STDMETHODCALLTYPE AddRef(
void);
699 virtual ULONG STDMETHODCALLTYPE Release(
void);
700 virtual HRESULT STDMETHODCALLTYPE VideoInputFormatChanged(BMDVideoInputFormatChangedEvents, IDeckLinkDisplayMode*, BMDDetectedVideoInputFormatFlags);
701 virtual HRESULT STDMETHODCALLTYPE VideoInputFrameArrived(IDeckLinkVideoInputFrame*, IDeckLinkAudioInputPacket*);
739 IDeckLinkAudioInputPacket *audioFrame,
741 int64_t abs_wallclock,
747 BMDTimeValue bmd_pts;
748 BMDTimeValue bmd_duration;
749 HRESULT res = E_INVALIDARG;
753 res = audioFrame->GetPacketTime(&bmd_pts, time_base.
den);
757 res = videoFrame->GetStreamTime(&bmd_pts, &bmd_duration, time_base.
den);
761 res = videoFrame->GetHardwareReferenceTimestamp(time_base.
den, &bmd_pts, &bmd_duration);
780 pts = bmd_pts / time_base.
num;
794 IDeckLinkTimecode *timecode;
796 #if BLACKMAGIC_DECKLINK_API_VERSION >= 0x0b000000 797 int hfr = (tc_format == bmdTimecodeRP188HighFrameRate);
801 if (videoFrame->GetTimecode(tc_format, &timecode) == S_OK) {
803 if (timecode->GetComponents(&hh, &mm, &ss, &ff) == S_OK) {
806 ff = ff << 1 | !!(timecode->GetFlags() & bmdTimecodeFieldMark);
823 #if BLACKMAGIC_DECKLINK_API_VERSION >= 0x0b000000 824 ret =
get_bmd_timecode(avctx, tc, frame_rate, bmdTimecodeRP188HighFrameRate, videoFrame);
827 ret =
get_bmd_timecode(avctx, tc, frame_rate, bmdTimecodeRP188VITC1, videoFrame);
829 ret =
get_bmd_timecode(avctx, tc, frame_rate, bmdTimecodeRP188VITC2, videoFrame);
831 ret =
get_bmd_timecode(avctx, tc, frame_rate, bmdTimecodeRP188LTC, videoFrame);
839 IDeckLinkVideoInputFrame *videoFrame, IDeckLinkAudioInputPacket *audioFrame)
842 void *audioFrameBytes;
843 BMDTimeValue frameTime;
844 BMDTimeValue frameDuration;
845 int64_t wallclock = 0, abs_wallclock = 0;
849 if (videoFrame && !(videoFrame->GetFlags() & bmdFrameHasNoInputSource) &&
861 if (
av_cmp_q(remainder, frame_duration) > 0) {
880 "Frame received (#%lu) - Valid (%liB) - QSize %fMB\n",
882 videoFrame->GetRowBytes() * videoFrame->GetHeight(),
883 (double)qsize / 1024 / 1024);
886 videoFrame->GetBytes(&frameBytes);
887 videoFrame->GetStreamTime(&frameTime, &frameDuration,
890 if (videoFrame->GetFlags() & bmdFrameHasNoInputSource) {
891 if (
ctx->
draw_bars && videoFrame->GetPixelFormat() == bmdFormat8BitYUV) {
893 0xEA80EA80, 0xD292D210, 0xA910A9A5, 0x90229035,
894 0x6ADD6ACA, 0x51EF515A, 0x286D28EF, 0x10801080 };
895 int width = videoFrame->GetWidth();
896 int height = videoFrame->GetHeight();
897 unsigned *p = (
unsigned *)frameBytes;
899 for (
int y = 0; y <
height; y++) {
900 for (
int x = 0; x <
width; x += 2)
901 *p++ = bars[(x * 8) /
width];
930 int size =
sizeof(uint32_t) * 4;
939 if (
av_dict_set(&metadata_dict,
"timecode", tc, 0) >= 0) {
942 if (packed_metadata) {
972 pkt.
size = videoFrame->GetRowBytes() *
973 videoFrame->GetHeight();
977 IDeckLinkVideoFrameAncillary *vanc;
986 if (videoFrame->GetAncillaryData(&vanc) == S_OK) {
988 BMDPixelFormat vanc_format = vanc->GetPixelFormat();
993 (vanc_format == bmdFormat8BitYUV || vanc_format == bmdFormat10BitYUV)) {
994 int64_t line_mask = 1;
996 for (i = 6; i < 336; i++, line_mask <<= 1) {
998 if ((
ctx->
teletext_lines & line_mask) && vanc->GetBufferForVerticalBlankingLine(i, (
void**)&buf) == S_OK) {
999 if (vanc_format == bmdFormat8BitYUV)
1000 txt_buf = teletext_data_unit_from_vbi_data(i, buf, txt_buf, VBI_PIXFMT_UYVY);
1002 txt_buf = teletext_data_unit_from_vbi_data_10bit(i, buf, txt_buf);
1009 if (vanc_format == bmdFormat10BitYUV && videoFrame->GetWidth() <=
MAX_WIDTH_VANC) {
1011 for (i = vanc_line_numbers[idx].vanc_start; i <= vanc_line_numbers[idx].
vanc_end; i++) {
1013 if (vanc->GetBufferForVerticalBlankingLine(i, (
void**)&buf) == S_OK) {
1015 size_t vanc_size = videoFrame->GetWidth();
1017 vanc_size = vanc_size * 2;
1023 txt_buf,
sizeof(txt_buf0) - (txt_buf - txt_buf0), &pkt);
1025 if (i == vanc_line_numbers[idx].field0_vanc_end)
1030 if (txt_buf - txt_buf0 > 1) {
1031 int stuffing_units = (4 - ((45 + txt_buf - txt_buf0) / 46) % 4) % 4;
1032 while (stuffing_units--) {
1033 memset(txt_buf, 0xff, 46);
1041 txt_pkt.
data = txt_buf0;
1042 txt_pkt.
size = txt_buf - txt_buf0;
1052 videoFrame->AddRef();
1062 BMDTimeValue audio_pts;
1067 audioFrame->GetBytes(&audioFrameBytes);
1086 BMDVideoInputFormatChangedEvents events, IDeckLinkDisplayMode *
mode,
1087 BMDDetectedVideoInputFormatFlags formatFlags)
1093 ctx->
raw_format = (formatFlags & bmdDetectedVideoInputRGB444) ? bmdFormat8BitARGB : bmdFormat8BitYUV;
1102 if (ctx->
attr->GetFlag(BMDDeckLinkSupportsInputFormatDetection, &autodetect_supported) != S_OK)
1104 if (autodetect_supported ==
false)
1111 bmdVideoInputEnableFormatDetection) != S_OK) {
1115 if (ctx->
dli->StartStreams() != S_OK) {
1120 for (
i = 0;
i < 30;
i++) {
1125 if (ctx->
bmd_mode != bmdModeUnknown &&
1130 ctx->
dli->PauseStreams();
1131 ctx->
dli->FlushStreams();
1133 if (ctx->
bmd_mode != bmdModeUnknown) {
1154 ctx->
dli->DisableVideoInput();
1155 ctx->
dli->DisableAudioInput();
1216 av_log(avctx,
AV_LOG_ERROR,
"Value for audio bit depth option must be either 16 or 32\n");
1231 if (
ctx->
dl->QueryInterface(IID_IDeckLinkInput, (
void **) &
ctx->
dli) != S_OK) {
1261 allocator->Release();
1286 av_log(avctx,
AV_LOG_ERROR,
"Libzvbi support is needed for capturing SD PAL teletext, please recompile FFmpeg.\n");
1301 st->codecpar->sample_rate = bmdAudioSampleRate48kHz;
1318 st->r_frame_rate =
av_make_q(st->time_base.den, st->time_base.num);
1321 case bmdFormat8BitYUV:
1326 case bmdFormat10BitYUV:
1330 case bmdFormat8BitARGB:
1335 case bmdFormat8BitBGRA:
1340 case bmdFormat10BitRGB:
1353 case bmdUpperFieldFirst:
1356 case bmdLowerFieldFirst:
1359 case bmdProgressiveFrame:
1360 case bmdProgressiveSegmentedFrame:
1409 bmdVideoInputFlagDefault);
1419 if (
ctx->
dli->StartStreams() != S_OK) {
1442 if (side_metadata) {
packed YUV 4:2:2, 16bpp, Cb Y0 Cr Y1
static int decklink_autodetect(struct decklink_cctx *cctx)
static int get_frame_timecode(AVFormatContext *avctx, decklink_ctx *ctx, AVTimecode *tc, IDeckLinkVideoInputFrame *videoFrame)
static int shift(int a, int b)
static av_always_inline int pthread_mutex_destroy(pthread_mutex_t *mutex)
static int64_t get_pkt_pts(IDeckLinkVideoInputFrame *videoFrame, IDeckLinkAudioInputPacket *audioFrame, int64_t wallclock, int64_t abs_wallclock, DecklinkPtsSource pts_src, AVRational time_base, int64_t *initial_pts, int copyts)
static const BMDPixelFormat decklink_raw_format_map[]
#define pthread_mutex_lock(a)
static av_always_inline int pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex)
ptrdiff_t const GLvoid * data
static uint8_t calc_parity_and_line_offset(int line)
#define AV_LOG_WARNING
Something somehow does not look correct.
const uint8_t ff_reverse[256]
ATSC A53 Part 4 Closed Captions.
int index
stream index in AVFormatContext
Convenience header that includes libavutil's core.
static void avpacket_queue_init(AVFormatContext *avctx, AVPacketQueue *q)
int ff_decklink_init_device(AVFormatContext *avctx, const char *name)
#define FF_ARRAY_ELEMS(a)
void * av_mallocz(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
int av_usleep(unsigned usec)
Sleep for a period of time.
static void error(const char *err)
static av_always_inline int pthread_cond_destroy(pthread_cond_t *cond)
virtual HRESULT STDMETHODCALLTYPE Decommit()
DecklinkPtsSource audio_pts_source
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static void handle_klv(AVFormatContext *avctx, decklink_ctx *ctx, IDeckLinkVideoInputFrame *videoFrame, int64_t pts)
static const BMDVideoConnection decklink_video_connection_map[]
const BMDDisplayMode AUTODETECT_DEFAULT_MODE
The exact code depends on how similar the blocks are and how related they are to the block
IDeckLinkProfileAttributes * attr
DecklinkPtsSource audio_pts_source
Opaque data information usually continuous.
int av_packet_unpack_dictionary(const uint8_t *data, int size, AVDictionary **dict)
Unpack a dictionary from side_data.
static void fill_data_unit_head(int line, uint8_t *tgt)
int64_t duration
Duration of this packet in AVStream->time_base units, 0 if unknown.
virtual ~decklink_allocator()
static const BMDAudioConnection decklink_audio_connection_map[]
AVStream * avformat_new_stream(AVFormatContext *s, const AVCodec *c)
Add a new stream to a media file.
void ff_decklink_list_devices_legacy(AVFormatContext *avctx, int show_inputs, int show_outputs)
AVDictionaryEntry * av_dict_get(const AVDictionary *m, const char *key, const AVDictionaryEntry *prev, int flags)
Get a dictionary entry with matching key.
s EdgeDetect Foobar g libavfilter vf_edgedetect c libavfilter vf_foobar c edit libavfilter and add an entry for foobar following the pattern of the other filters edit libavfilter allfilters and add an entry for foobar following the pattern of the other filters configure make j< whatever > ffmpeg ffmpeg i you should get a foobar png with Lena edge detected That s it
void av_packet_move_ref(AVPacket *dst, AVPacket *src)
Move every field in src to dst and reset src.
#define AV_LOG_VERBOSE
Detailed information.
static av_always_inline int pthread_cond_signal(pthread_cond_t *cond)
static uint8_t * vanc_to_cc(AVFormatContext *avctx, uint16_t *buf, size_t words, unsigned &cc_count)
int ff_decklink_list_devices(AVFormatContext *avctx, struct AVDeviceInfoList *device_list, int show_inputs, int show_outputs)
static int get_vanc_line_idx(BMDDisplayMode mode)
#define AV_PKT_FLAG_KEY
The packet contains a keyframe.
int64_t av_rescale_q(int64_t a, AVRational bq, AVRational cq)
Rescale a 64-bit integer by 2 rational numbers.
BMDVideoConnection video_input
Main libavdevice API header.
virtual ULONG STDMETHODCALLTYPE Release(void)
static void avpacket_queue_flush(AVPacketQueue *q)
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
static const uint16_t mask[17]
static int get_bmd_timecode(AVFormatContext *avctx, AVTimecode *tc, AVRational frame_rate, BMDTimecodeFormat tc_format, IDeckLinkVideoInputFrame *videoFrame)
uint8_t * av_packet_get_side_data(const AVPacket *pkt, enum AVPacketSideDataType type, int *size)
Get side information from packet.
virtual ULONG STDMETHODCALLTYPE AddRef(void)
char * url
input or output URL.
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
void av_dict_free(AVDictionary **pm)
Free all the memory allocated for an AVDictionary struct and all keys and values. ...
#define AV_FOURCC_MAX_STRING_SIZE
AVBufferRef * buf
A reference to the reference-counted buffer where the packet data is stored.
simple assert() macros that are a bit more flexible than ISO C assert().
AVBufferRef * av_buffer_create(uint8_t *data, int size, void(*free)(void *opaque, uint8_t *data), void *opaque, int flags)
Create an AVBuffer from an existing array.
static uint8_t * teletext_data_unit_from_ancillary_packet(uint16_t *py, uint16_t *pend, uint8_t *tgt, int64_t wanted_lines, int allow_multipacket)
av_cold int ff_decklink_read_close(AVFormatContext *avctx)
int flags
A combination of AV_PKT_FLAG values.
int ff_decklink_set_configs(AVFormatContext *avctx, decklink_direction_t direction)
#define ss(width, name, subs,...)
int64_t av_rescale(int64_t a, int64_t b, int64_t c)
Rescale a 64-bit integer with rounding to nearest.
#define AV_TIME_BASE
Internal time base represented as integer.
int av_packet_make_refcounted(AVPacket *pkt)
Ensure the data described by a given packet is reference counted.
virtual HRESULT STDMETHODCALLTYPE Commit()
static av_always_inline int pthread_mutex_init(pthread_mutex_t *mutex, const pthread_mutexattr_t *attr)
int ff_decklink_list_input_devices(AVFormatContext *avctx, struct AVDeviceInfoList *device_list)
int ff_decklink_list_formats(AVFormatContext *avctx, decklink_direction_t direction)
#define pthread_mutex_unlock(a)
static volatile int checksum
static int avpacket_queue_put(AVPacketQueue *q, AVPacket *pkt)
#define AVERROR_EXIT
Immediate exit was requested; the called function should not be restarted.
static void extract_luma_from_v210(uint16_t *dst, const uint8_t *src, int width)
BMDPixelFormat raw_format
char * av_fourcc_make_string(char *buf, uint32_t fourcc)
Fill the provided buffer with a string containing a FourCC (four-character code) representation.
virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv)
static uint8_t * get_metadata(AVFormatContext *avctx, uint16_t *buf, size_t width, uint8_t *tgt, size_t tgt_size, AVPacket *pkt)
static const BMDTimecodeFormat decklink_timecode_format_map[]
int64_t av_gettime(void)
Get the current time in microseconds.
int ff_decklink_set_format(AVFormatContext *avctx, int width, int height, int tb_num, int tb_den, enum AVFieldOrder field_order, decklink_direction_t direction)
#define AV_LOG_INFO
Standard information.
A list of zero terminated key/value strings.
static AVRational av_make_q(int num, int den)
Create an AVRational.
uint8_t * av_packet_pack_dictionary(AVDictionary *dict, int *size)
Pack a dictionary for use in side_data.
DecklinkPtsSource video_pts_source
void av_packet_unref(AVPacket *pkt)
Wipe the packet.
Tag MUST be and< 10hcoeff half pel interpolation filter coefficients, hcoeff[0] are the 2 middle coefficients[1] are the next outer ones and so on, resulting in a filter like:...eff[2], hcoeff[1], hcoeff[0], hcoeff[0], hcoeff[1], hcoeff[2]...the sign of the coefficients is not explicitly stored but alternates after each coeff and coeff[0] is positive, so...,+,-,+,-,+,+,-,+,-,+,...hcoeff[0] is not explicitly stored but found by subtracting the sum of all stored coefficients with signs from 32 hcoeff[0]=32-hcoeff[1]-hcoeff[2]-...a good choice for hcoeff and htaps is htaps=6 hcoeff={40,-10, 2}an alternative which requires more computations at both encoder and decoder side and may or may not be better is htaps=8 hcoeff={42,-14, 6,-2}ref_frames minimum of the number of available reference frames and max_ref_frames for example the first frame after a key frame always has ref_frames=1spatial_decomposition_type wavelet type 0 is a 9/7 symmetric compact integer wavelet 1 is a 5/3 symmetric compact integer wavelet others are reserved stored as delta from last, last is reset to 0 if always_reset||keyframeqlog quality(logarithmic quantizer scale) stored as delta from last, last is reset to 0 if always_reset||keyframemv_scale stored as delta from last, last is reset to 0 if always_reset||keyframe FIXME check that everything works fine if this changes between framesqbias dequantization bias stored as delta from last, last is reset to 0 if always_reset||keyframeblock_max_depth maximum depth of the block tree stored as delta from last, last is reset to 0 if always_reset||keyframequant_table quantization tableHighlevel bitstream structure:==============================--------------------------------------------|Header|--------------------------------------------|------------------------------------|||Block0||||split?||||yes no||||.........intra?||||:Block01:yes no||||:Block02:.................||||:Block03::y DC::ref index:||||:Block04::cb DC::motion x:||||.........:cr DC::motion y:||||.................|||------------------------------------||------------------------------------|||Block1|||...|--------------------------------------------|------------------------------------|||Y subbands||Cb subbands||Cr subbands||||------||------||------|||||LL0||HL0||||LL0||HL0||||LL0||HL0|||||------||------||------||||------||------||------|||||LH0||HH0||||LH0||HH0||||LH0||HH0|||||------||------||------||||------||------||------|||||HL1||LH1||||HL1||LH1||||HL1||LH1|||||------||------||------||||------||------||------|||||HH1||HL2||||HH1||HL2||||HH1||HL2|||||...||...||...|||------------------------------------|--------------------------------------------Decoding process:=================------------|||Subbands|------------||||------------|Intra DC||||LL0 subband prediction------------|\Dequantization-------------------\||Reference frames|\IDWT|--------------|Motion\|||Frame 0||Frame 1||Compensation.OBMC v-------|--------------|--------------.\------> Frame n output Frame Frame<----------------------------------/|...|-------------------Range Coder:============Binary Range Coder:-------------------The implemented range coder is an adapted version based upon"Range encoding: an algorithm for removing redundancy from a digitised message."by G.N.N.Martin.The symbols encoded by the Snow range coder are bits(0|1).The associated probabilities are not fix but change depending on the symbol mix seen so far.bit seen|new state---------+-----------------------------------------------0|256-state_transition_table[256-old_state];1|state_transition_table[old_state];state_transition_table={0, 0, 0, 0, 0, 0, 0, 0, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 190, 191, 192, 194, 194, 195, 196, 197, 198, 199, 200, 201, 202, 202, 204, 205, 206, 207, 208, 209, 209, 210, 211, 212, 213, 215, 215, 216, 217, 218, 219, 220, 220, 222, 223, 224, 225, 226, 227, 227, 229, 229, 230, 231, 232, 234, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 248, 0, 0, 0, 0, 0, 0, 0};FIXME Range Coding of integers:-------------------------FIXME Neighboring Blocks:===================left and top are set to the respective blocks unless they are outside of the image in which case they are set to the Null block top-left is set to the top left block unless it is outside of the image in which case it is set to the left block if this block has no larger parent block or it is at the left side of its parent block and the top right block is not outside of the image then the top right block is used for top-right else the top-left block is used Null block y, cb, cr are 128 level, ref, mx and my are 0 Motion Vector Prediction:=========================1.the motion vectors of all the neighboring blocks are scaled to compensate for the difference of reference frames scaled_mv=(mv *(256 *(current_reference+1)/(mv.reference+1))+128)> the median of the scaled top and top right vectors is used as motion vector prediction the used motion vector is the sum of the predictor and(mvx_diff, mvy_diff)*mv_scale Intra DC Prediction block[y][x] dc[1]
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining list
static void decklink_object_free(void *opaque, uint8_t *data)
static void input_callback(MMAL_PORT_T *port, MMAL_BUFFER_HEADER_T *buffer)
virtual HRESULT STDMETHODCALLTYPE AllocateBuffer(unsigned int bufferSize, void **allocatedBuffer)
int av_dict_set(AVDictionary **pm, const char *key, const char *value, int flags)
Set the given entry in *pm, overwriting an existing entry.
static uint8_t * teletext_data_unit_from_op47_vbi_packet(int line, uint16_t *py, uint8_t *tgt)
static int avpacket_queue_get(AVPacketQueue *q, AVPacket *pkt, int block)
Rational number (pair of numerator and denominator).
static int linemask_matches(int line, int64_t mask)
FF_ENABLE_DEPRECATION_WARNINGS int av_packet_add_side_data(AVPacket *pkt, enum AVPacketSideDataType type, uint8_t *data, size_t size)
Wrap an existing array as a packet side data.
BMDTimecodeFormat tc_format
packed BGR 8:8:8, 32bpp, BGRXBGRX... X=unused/undefined
int64_t av_gettime_relative(void)
Get the current time in microseconds since some unspecified starting point.
#define flags(name, subs,...)
void ff_decklink_cleanup(AVFormatContext *avctx)
char * av_timecode_make_string(const AVTimecode *tc, char *buf, int framenum)
Load timecode string in buf.
Timecode which conforms to SMPTE ST 12-1:2014.
static av_always_inline AVRational av_inv_q(AVRational q)
Invert a rational.
static int av_cmp_q(AVRational a, AVRational b)
Compare two rationals.
static int check_vanc_parity_checksum(uint16_t *buf, int len, uint16_t checksum)
BMDAudioConnection audio_input
int av_timecode_init_from_components(AVTimecode *tc, AVRational rate, int flags, int hh, int mm, int ss, int ff, void *log_ctx)
Init a timecode struct from the passed timecode components.
The official guide to swscale for confused that consecutive non overlapping rectangles of slice_bottom special converter These generally are unscaled converters of common like for each output line the vertical scaler pulls lines from a ring buffer When the ring buffer does not contain the wanted line
static void avpacket_queue_end(AVPacketQueue *q)
struct AVPacketList * next
common internal and external API header
static unsigned long long avpacket_queue_size(AVPacketQueue *q)
int ff_decklink_read_packet(AVFormatContext *avctx, AVPacket *pkt)
static av_always_inline int pthread_cond_init(pthread_cond_t *cond, const pthread_condattr_t *attr)
void av_init_packet(AVPacket *pkt)
Initialize optional fields of a packet with default values.
#define AV_INPUT_BUFFER_PADDING_SIZE
Required number of additionally allocated bytes at the end of the input bitstream for decoding...
static void clear_parity_bits(uint16_t *buf, int len)
void * priv_data
Format private data.
virtual HRESULT STDMETHODCALLTYPE ReleaseBuffer(void *buffer)
av_cold int ff_decklink_read_header(AVFormatContext *avctx)
static void unpack_v210(uint16_t *dst, const uint8_t *src, int width)
static VANCLineNumber vanc_line_numbers[]
int64_t dts
Decompression timestamp in AVStream->time_base units; the time at which the packet is decompressed...
and forward the result(frame or status change) to the corresponding input.If nothing is possible
DecklinkPtsSource video_pts_source
AVCodecParameters * codecpar
Codec parameters associated with this stream.
static uint8_t * teletext_data_unit_from_op47_data(uint16_t *py, uint16_t *pend, uint8_t *tgt, int64_t wanted_lines)
uint8_t * av_packet_new_side_data(AVPacket *pkt, enum AVPacketSideDataType type, int size)
Allocate new information of a packet.
AVRational time_base
This is the fundamental unit of time (in seconds) in terms of which frame timestamps are represented...
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
packed RGB 8:8:8, 32bpp, XRGBXRGB... X=unused/undefined
AVRational r_frame_rate
Real base framerate of the stream.
#define AVERROR_EXTERNAL
Generic error in an external library.
This structure stores compressed data.
mode
Use these values in ebur128_init (or'ed).
int64_t pts
Presentation timestamp in AVStream->time_base units; the time at which the decompressed packet will b...
#define AV_NOPTS_VALUE
Undefined timestamp value.
#define AV_TIMECODE_STR_SIZE
uint32_t av_timecode_get_smpte_from_framenum(const AVTimecode *tc, int framenum)
Convert frame number to SMPTE 12M binary representation.