FFmpeg
notchlc.c
Go to the documentation of this file.
1 /*
2  * NotchLC decoder
3  * Copyright (c) 2020 Paul B Mahol
4  *
5  * This file is part of FFmpeg.
6  *
7  * FFmpeg is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * FFmpeg is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with FFmpeg; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21 
22 #include <stdio.h>
23 #include <stdlib.h>
24 #include <string.h>
25 
26 #define BITSTREAM_READER_LE
27 #include "libavutil/intreadwrite.h"
28 #include "avcodec.h"
29 #include "bytestream.h"
30 #include "get_bits.h"
31 #include "internal.h"
32 #include "lzf.h"
33 #include "thread.h"
34 
35 typedef struct NotchLCContext {
36  unsigned compressed_size;
37  unsigned format;
38 
41 
43  int64_t lzf_size;
44 
45  unsigned texture_size_x;
46  unsigned texture_size_y;
51  unsigned y_data_offset;
52  unsigned uv_data_offset;
53  unsigned y_data_size;
54  unsigned a_data_offset;
55  unsigned uv_count_offset;
56  unsigned a_count_size;
57  unsigned data_end;
58 
62 
64 {
67  avctx->colorspace = AVCOL_SPC_RGB;
70 
71  return 0;
72 }
73 
74 #define HISTORY_SIZE (64 * 1024)
75 
76 static int lz4_decompress(AVCodecContext *avctx,
79 {
80  unsigned reference_pos, match_length, delta, pos = 0;
81  uint8_t history[64 * 1024];
82 
83  while (bytestream2_get_bytes_left(gb) > 0) {
84  uint8_t token = bytestream2_get_byte(gb);
85  unsigned num_literals = token >> 4;
86 
87  if (num_literals == 15) {
88  unsigned char current;
89  do {
90  current = bytestream2_get_byte(gb);
91  num_literals += current;
92  } while (current == 255);
93  }
94 
95  if (pos + num_literals < HISTORY_SIZE) {
96  bytestream2_get_buffer(gb, history + pos, num_literals);
97  pos += num_literals;
98  } else {
99  while (num_literals-- > 0) {
100  history[pos++] = bytestream2_get_byte(gb);
101  if (pos == HISTORY_SIZE) {
102  bytestream2_put_buffer(pb, history, HISTORY_SIZE);
103  pos = 0;
104  }
105  }
106  }
107 
108  if (bytestream2_get_bytes_left(gb) <= 0)
109  break;
110 
111  delta = bytestream2_get_le16(gb);
112  if (delta == 0)
113  return 0;
114  match_length = 4 + (token & 0x0F);
115  if (match_length == 4 + 0x0F) {
116  uint8_t current;
117 
118  do {
119  current = bytestream2_get_byte(gb);
120  match_length += current;
121  } while (current == 255);
122  }
123  reference_pos = (pos >= delta) ? (pos - delta) : (HISTORY_SIZE + pos - delta);
124  if (pos + match_length < HISTORY_SIZE && reference_pos + match_length < HISTORY_SIZE) {
125  if (pos >= reference_pos + match_length || reference_pos >= pos + match_length) {
126  memcpy(history + pos, history + reference_pos, match_length);
127  pos += match_length;
128  } else {
129  while (match_length-- > 0)
130  history[pos++] = history[reference_pos++];
131  }
132  } else {
133  while (match_length-- > 0) {
134  history[pos++] = history[reference_pos++];
135  if (pos == HISTORY_SIZE) {
136  bytestream2_put_buffer(pb, history, HISTORY_SIZE);
137  pos = 0;
138  }
139  reference_pos %= HISTORY_SIZE;
140  }
141  }
142  }
143 
144  bytestream2_put_buffer(pb, history, pos);
145 
146  return bytestream2_tell_p(pb);
147 }
148 
150  unsigned uncompressed_size)
151 {
152  NotchLCContext *s = avctx->priv_data;
153  GetByteContext rgb, dgb, *gb = &s->gb;
155  int ylinesize, ulinesize, vlinesize, alinesize;
156  uint16_t *dsty, *dstu, *dstv, *dsta;
157  int ret;
158 
159  s->texture_size_x = bytestream2_get_le32(gb);
160  s->texture_size_y = bytestream2_get_le32(gb);
161 
162  ret = ff_set_dimensions(avctx, s->texture_size_x, s->texture_size_y);
163  if (ret < 0)
164  return ret;
165 
166  s->uv_offset_data_offset = bytestream2_get_le32(gb);
167  if (s->uv_offset_data_offset >= UINT_MAX / 4)
168  return AVERROR_INVALIDDATA;
169  s->uv_offset_data_offset *= 4;
170  if (s->uv_offset_data_offset >= uncompressed_size)
171  return AVERROR_INVALIDDATA;
172 
173  s->y_control_data_offset = bytestream2_get_le32(gb);
174  if (s->y_control_data_offset >= UINT_MAX / 4)
175  return AVERROR_INVALIDDATA;
176  s->y_control_data_offset *= 4;
177  if (s->y_control_data_offset >= uncompressed_size)
178  return AVERROR_INVALIDDATA;
179 
180  s->a_control_word_offset = bytestream2_get_le32(gb);
181  if (s->a_control_word_offset >= UINT_MAX / 4)
182  return AVERROR_INVALIDDATA;
183  s->a_control_word_offset *= 4;
184  if (s->a_control_word_offset >= uncompressed_size)
185  return AVERROR_INVALIDDATA;
186 
187  s->uv_data_offset = bytestream2_get_le32(gb);
188  if (s->uv_data_offset >= UINT_MAX / 4)
189  return AVERROR_INVALIDDATA;
190  s->uv_data_offset *= 4;
191  if (s->uv_data_offset >= uncompressed_size)
192  return AVERROR_INVALIDDATA;
193 
194  s->y_data_size = bytestream2_get_le32(gb);
195  if (s->y_data_size >= UINT_MAX / 4)
196  return AVERROR_INVALIDDATA;
197 
198  s->a_data_offset = bytestream2_get_le32(gb);
199  if (s->a_data_offset >= UINT_MAX / 4)
200  return AVERROR_INVALIDDATA;
201  s->a_data_offset *= 4;
202  if (s->a_data_offset >= uncompressed_size)
203  return AVERROR_INVALIDDATA;
204 
205  s->a_count_size = bytestream2_get_le32(gb);
206  if (s->a_count_size >= UINT_MAX / 4)
207  return AVERROR_INVALIDDATA;
208  s->a_count_size *= 4;
209  if (s->a_count_size >= uncompressed_size)
210  return AVERROR_INVALIDDATA;
211 
212  s->data_end = bytestream2_get_le32(gb);
213  if (s->data_end > uncompressed_size)
214  return AVERROR_INVALIDDATA;
215 
217  if (s->data_end <= s->y_data_size)
218  return AVERROR_INVALIDDATA;
219  s->y_data_offset = s->data_end - s->y_data_size;
220  if (s->y_data_offset <= s->a_data_offset)
221  return AVERROR_INVALIDDATA;
223 
224  if ((ret = ff_thread_get_buffer(avctx, frame, 0)) < 0)
225  return ret;
226 
227  rgb = *gb;
228  dgb = *gb;
229  bytestream2_seek(&rgb, s->y_data_row_offsets, SEEK_SET);
230  bytestream2_seek(gb, s->y_control_data_offset, SEEK_SET);
231 
232  dsty = (uint16_t *)p->data[0];
233  dsta = (uint16_t *)p->data[3];
234  ylinesize = p->linesize[0] / 2;
235  alinesize = p->linesize[3] / 2;
236 
237  for (int y = 0; y < avctx->height; y += 4) {
238  const unsigned row_offset = bytestream2_get_le32(&rgb);
239 
240  bytestream2_seek(&dgb, s->y_data_offset + row_offset, SEEK_SET);
241 
243  for (int x = 0; x < avctx->width; x += 4) {
244  unsigned item = bytestream2_get_le32(gb);
245  unsigned y_min = item & 4095;
246  unsigned y_max = (item >> 12) & 4095;
247  unsigned y_diff = y_max - y_min;
248  unsigned control[4];
249 
250  control[0] = (item >> 24) & 3;
251  control[1] = (item >> 26) & 3;
252  control[2] = (item >> 28) & 3;
253  control[3] = (item >> 30) & 3;
254 
255  for (int i = 0; i < 4; i++) {
256  const int nb_bits = control[i] + 1;
257  const int div = (1 << nb_bits) - 1;
258  const int add = div - 1;
259 
260  dsty[x + i * ylinesize + 0] = av_clip_uintp2(y_min + ((y_diff * get_bits(&bit, nb_bits) + add) / div), 12);
261  dsty[x + i * ylinesize + 1] = av_clip_uintp2(y_min + ((y_diff * get_bits(&bit, nb_bits) + add) / div), 12);
262  dsty[x + i * ylinesize + 2] = av_clip_uintp2(y_min + ((y_diff * get_bits(&bit, nb_bits) + add) / div), 12);
263  dsty[x + i * ylinesize + 3] = av_clip_uintp2(y_min + ((y_diff * get_bits(&bit, nb_bits) + add) / div), 12);
264  }
265  }
266 
267  dsty += 4 * ylinesize;
268  }
269 
270  rgb = *gb;
271  dgb = *gb;
272  bytestream2_seek(gb, s->a_control_word_offset, SEEK_SET);
273  if (s->uv_count_offset == s->a_control_word_offset) {
274  for (int y = 0; y < avctx->height; y++) {
275  for (int x = 0; x < avctx->width; x++)
276  dsta[x] = 4095;
277  dsta += alinesize;
278  }
279  } else {
280  for (int y = 0; y < avctx->height; y += 16) {
281  for (int x = 0; x < avctx->width; x += 16) {
282  unsigned m = bytestream2_get_le32(gb);
283  unsigned offset = bytestream2_get_le32(gb);
284  unsigned alpha0, alpha1;
285  uint64_t control;
286 
287  if (offset >= UINT_MAX / 4)
288  return AVERROR_INVALIDDATA;
289  offset = offset * 4 + s->uv_data_offset + s->a_data_offset;
290  if (offset >= s->data_end)
291  return AVERROR_INVALIDDATA;
292 
293  bytestream2_seek(&dgb, offset, SEEK_SET);
294  control = bytestream2_get_le64(&dgb);
295  alpha0 = control & 0xFF;
296  alpha1 = (control >> 8) & 0xFF;
297  control = control >> 16;
298 
299  for (int by = 0; by < 4; by++) {
300  for (int bx = 0; bx < 4; bx++) {
301  switch (m & 3) {
302  case 0:
303  for (int i = 0; i < 4; i++) {
304  for (int j = 0; j < 4; j++) {
305  dsta[x + (i + by * 4) * alinesize + bx * 4 + j] = 0;
306  }
307  }
308  break;
309  case 1:
310  for (int i = 0; i < 4; i++) {
311  for (int j = 0; j < 4; j++) {
312  dsta[x + (i + by * 4) * alinesize + bx * 4 + j] = 4095;
313  }
314  }
315  break;
316  case 2:
317  for (int i = 0; i < 4; i++) {
318  for (int j = 0; j < 4; j++) {
319  dsta[x + (i + by * 4) * alinesize + bx * 4 + j] = (alpha0 + (alpha1 - alpha0) * (control & 7)) << 4;
320  }
321  }
322  break;
323  default:
324  return AVERROR_INVALIDDATA;
325  }
326 
327  control >>= 3;
328  m >>= 2;
329  }
330  }
331  }
332 
333  dsta += 16 * alinesize;
334  }
335  }
336 
337  bytestream2_seek(&rgb, s->uv_offset_data_offset, SEEK_SET);
338 
339  dstu = (uint16_t *)p->data[1];
340  dstv = (uint16_t *)p->data[2];
341  ulinesize = p->linesize[1] / 2;
342  vlinesize = p->linesize[2] / 2;
343 
344  for (int y = 0; y < avctx->height; y += 16) {
345  for (int x = 0; x < avctx->width; x += 16) {
346  unsigned offset = bytestream2_get_le32(&rgb) * 4;
347  int u[16][16] = { 0 }, v[16][16] = { 0 };
348  int u0, v0, u1, v1, udif, vdif;
349  unsigned escape, is8x8, loc;
350 
351  bytestream2_seek(&dgb, s->uv_data_offset + offset, SEEK_SET);
352 
353  is8x8 = bytestream2_get_le16(&dgb);
354  escape = bytestream2_get_le16(&dgb);
355 
356  if (escape == 0 && is8x8 == 0) {
357  u0 = bytestream2_get_byte(&dgb);
358  v0 = bytestream2_get_byte(&dgb);
359  u1 = bytestream2_get_byte(&dgb);
360  v1 = bytestream2_get_byte(&dgb);
361  loc = bytestream2_get_le32(&dgb);
362  u0 = (u0 << 4) | (u0 & 0xF);
363  v0 = (v0 << 4) | (v0 & 0xF);
364  u1 = (u1 << 4) | (u1 & 0xF);
365  v1 = (v1 << 4) | (v1 & 0xF);
366  udif = u1 - u0;
367  vdif = v1 - v0;
368 
369  for (int i = 0; i < 16; i += 4) {
370  for (int j = 0; j < 16; j += 4) {
371  for (int ii = 0; ii < 4; ii++) {
372  for (int jj = 0; jj < 4; jj++) {
373  u[i + ii][j + jj] = u0 + ((udif * (int)(loc & 3) + 2) / 3);
374  v[i + ii][j + jj] = v0 + ((vdif * (int)(loc & 3) + 2) / 3);
375  }
376  }
377 
378  loc >>= 2;
379  }
380  }
381  } else {
382  for (int i = 0; i < 16; i += 8) {
383  for (int j = 0; j < 16; j += 8) {
384  if (is8x8 & 1) {
385  u0 = bytestream2_get_byte(&dgb);
386  v0 = bytestream2_get_byte(&dgb);
387  u1 = bytestream2_get_byte(&dgb);
388  v1 = bytestream2_get_byte(&dgb);
389  loc = bytestream2_get_le32(&dgb);
390  u0 = (u0 << 4) | (u0 & 0xF);
391  v0 = (v0 << 4) | (v0 & 0xF);
392  u1 = (u1 << 4) | (u1 & 0xF);
393  v1 = (v1 << 4) | (v1 & 0xF);
394  udif = u1 - u0;
395  vdif = v1 - v0;
396 
397  for (int ii = 0; ii < 8; ii += 2) {
398  for (int jj = 0; jj < 8; jj += 2) {
399  for (int iii = 0; iii < 2; iii++) {
400  for (int jjj = 0; jjj < 2; jjj++) {
401  u[i + ii + iii][j + jj + jjj] = u0 + ((udif * (int)(loc & 3) + 2) / 3);
402  v[i + ii + iii][j + jj + jjj] = v0 + ((vdif * (int)(loc & 3) + 2) / 3);
403  }
404  }
405 
406  loc >>= 2;
407  }
408  }
409  } else if (escape) {
410  for (int ii = 0; ii < 8; ii += 4) {
411  for (int jj = 0; jj < 8; jj += 4) {
412  u0 = bytestream2_get_byte(&dgb);
413  v0 = bytestream2_get_byte(&dgb);
414  u1 = bytestream2_get_byte(&dgb);
415  v1 = bytestream2_get_byte(&dgb);
416  loc = bytestream2_get_le32(&dgb);
417  u0 = (u0 << 4) | (u0 & 0xF);
418  v0 = (v0 << 4) | (v0 & 0xF);
419  u1 = (u1 << 4) | (u1 & 0xF);
420  v1 = (v1 << 4) | (v1 & 0xF);
421  udif = u1 - u0;
422  vdif = v1 - v0;
423 
424  for (int iii = 0; iii < 4; iii++) {
425  for (int jjj = 0; jjj < 4; jjj++) {
426  u[i + ii + iii][j + jj + jjj] = u0 + ((udif * (int)(loc & 3) + 2) / 3);
427  v[i + ii + iii][j + jj + jjj] = v0 + ((vdif * (int)(loc & 3) + 2) / 3);
428 
429  loc >>= 2;
430  }
431  }
432  }
433  }
434  }
435 
436  is8x8 >>= 1;
437  }
438  }
439  }
440 
441  for (int i = 0; i < 16; i++) {
442  for (int j = 0; j < 16; j++) {
443  dstu[x + i * ulinesize + j] = u[i][j];
444  dstv[x + i * vlinesize + j] = v[i][j];
445  }
446  }
447  }
448 
449  dstu += 16 * ulinesize;
450  dstv += 16 * vlinesize;
451  }
452 
453  return 0;
454 }
455 
456 static int decode_frame(AVCodecContext *avctx,
457  void *data, int *got_frame,
458  AVPacket *avpkt)
459 {
460  NotchLCContext *s = avctx->priv_data;
461  ThreadFrame frame = { .f = data };
462  GetByteContext *gb = &s->gb;
463  PutByteContext *pb = &s->pb;
464  unsigned uncompressed_size;
465  AVFrame *p = data;
466  int ret;
467 
468  if (avpkt->size <= 40)
469  return AVERROR_INVALIDDATA;
470 
471  bytestream2_init(gb, avpkt->data, avpkt->size);
472 
473  if (bytestream2_get_le32(gb) != MKBETAG('N','L','C','1'))
474  return AVERROR_INVALIDDATA;
475 
476  uncompressed_size = bytestream2_get_le32(gb);
477  s->compressed_size = bytestream2_get_le32(gb);
478  s->format = bytestream2_get_le32(gb);
479 
480  if (s->format > 2)
481  return AVERROR_PATCHWELCOME;
482 
483  if (s->format == 0) {
484  ret = ff_lzf_uncompress(gb, &s->lzf_buffer, &s->lzf_size);
485  if (ret < 0)
486  return ret;
487 
488  if (uncompressed_size > s->lzf_size)
489  return AVERROR_INVALIDDATA;
490 
491  bytestream2_init(gb, s->lzf_buffer, uncompressed_size);
492  } else if (s->format == 1) {
494  uncompressed_size);
495  if (!s->uncompressed_buffer)
496  return AVERROR(ENOMEM);
497 
499 
500  ret = lz4_decompress(avctx, gb, pb);
501  if (ret != uncompressed_size)
502  return AVERROR_INVALIDDATA;
503 
504  bytestream2_init(gb, s->uncompressed_buffer, uncompressed_size);
505  }
506 
507  ret = decode_blocks(avctx, p, &frame, uncompressed_size);
508  if (ret < 0)
509  return ret;
510 
512  p->key_frame = 1;
513 
514  *got_frame = 1;
515 
516  return avpkt->size;
517 }
518 
520 {
521  NotchLCContext *s = avctx->priv_data;
522 
524  s->uncompressed_size = 0;
525  av_freep(&s->lzf_buffer);
526  s->lzf_size = 0;
527 
528  return 0;
529 }
530 
532  .name = "notchlc",
533  .long_name = NULL_IF_CONFIG_SMALL("NotchLC"),
534  .type = AVMEDIA_TYPE_VIDEO,
535  .id = AV_CODEC_ID_NOTCHLC,
536  .priv_data_size = sizeof(NotchLCContext),
537  .init = decode_init,
538  .close = decode_end,
539  .decode = decode_frame,
541 };
int64_t lzf_size
Definition: notchlc.c:43
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:59
This structure describes decoded (raw) audio or video data.
Definition: frame.h:308
unsigned a_control_word_offset
Definition: notchlc.c:50
ptrdiff_t const GLvoid * data
Definition: opengl_enc.c:100
static unsigned int get_bits(GetBitContext *s, int n)
Read 1-25 bits.
Definition: get_bits.h:379
unsigned uncompressed_size
Definition: notchlc.c:40
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
Definition: utils.c:106
static av_cold int init(AVCodecContext *avctx)
Definition: avrndec.c:35
unsigned uv_count_offset
Definition: notchlc.c:55
enum AVColorRange color_range
MPEG vs JPEG YUV range.
Definition: avcodec.h:1161
uint8_t * lzf_buffer
Definition: notchlc.c:42
int size
Definition: packet.h:364
static av_always_inline void bytestream2_init_writer(PutByteContext *p, uint8_t *buf, int buf_size)
Definition: bytestream.h:143
unsigned y_data_row_offsets
Definition: notchlc.c:47
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:736
static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size)
Definition: bytestream.h:133
void av_fast_padded_malloc(void *ptr, unsigned int *size, size_t min_size)
Same behaviour av_fast_malloc but the buffer has additional AV_INPUT_BUFFER_PADDING_SIZE at the end w...
Definition: utils.c:72
static av_cold int decode_init(AVCodecContext *avctx)
Definition: notchlc.c:63
GLfloat v0
Definition: opengl_enc.c:106
AVCodec.
Definition: codec.h:190
order of coefficients is actually GBR, also IEC 61966-2-1 (sRGB)
Definition: pixfmt.h:513
static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, FILE *outfile)
Definition: decode_audio.c:71
unsigned uv_data_offset
Definition: notchlc.c:52
PutByteContext pb
Definition: notchlc.c:60
unsigned y_data_size
Definition: notchlc.c:53
uint8_t
#define av_cold
Definition: attributes.h:88
float delta
AVCodec ff_notchlc_decoder
Definition: notchlc.c:531
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
Multithreading support functions.
static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt)
Definition: notchlc.c:456
#define u(width, name, range_min, range_max)
Definition: cbs_h2645.c:262
uint8_t * data
Definition: packet.h:363
const uint8_t * buffer
Definition: bytestream.h:34
unsigned y_control_data_offset
Definition: notchlc.c:49
bitstream reader API header.
static av_always_inline int bytestream2_get_bytes_left(GetByteContext *g)
Definition: bytestream.h:154
int ff_lzf_uncompress(GetByteContext *gb, uint8_t **buf, int64_t *size)
Definition: lzf.c:40
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
Definition: internal.h:153
static av_always_inline unsigned int bytestream2_get_buffer(GetByteContext *g, uint8_t *dst, unsigned int size)
Definition: bytestream.h:263
unsigned int pos
Definition: spdifenc.c:410
unsigned format
Definition: notchlc.c:37
also ITU-R BT1361 / IEC 61966-2-4 / SMPTE RP177 Annex B
Definition: pixfmt.h:460
const char * name
Name of the codec implementation.
Definition: codec.h:197
static av_always_inline int bytestream2_tell_p(PutByteContext *p)
Definition: bytestream.h:193
#define AV_CODEC_CAP_FRAME_THREADS
Codec supports frame-level multithreading.
Definition: codec.h:106
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before as well as code calling up to before the decode process starts Call have so the codec calls ff_thread_report set FF_CODEC_CAP_ALLOCATE_PROGRESS in AVCodec caps_internal and use ff_thread_get_buffer() to allocate frames.The frames must then be freed with ff_thread_release_buffer().Otherwise decode directly into the user-supplied frames.Call ff_thread_report_progress() after some part of the current picture has decoded.A good place to put this is where draw_horiz_band() is called-add this if it isn't called anywhere
enum AVPictureType pict_type
Picture type of the frame.
Definition: frame.h:391
static int decode_blocks(AVCodecContext *avctx, AVFrame *p, ThreadFrame *frame, unsigned uncompressed_size)
Definition: notchlc.c:149
#define AV_PIX_FMT_YUVA444P12
Definition: pixfmt.h:440
int width
picture width / height.
Definition: avcodec.h:699
enum AVColorPrimaries color_primaries
Chromaticity coordinates of the source primaries.
Definition: avcodec.h:1140
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
#define s(width, name)
Definition: cbs_vp9.c:257
GetByteContext gb
Definition: notchlc.c:59
static av_always_inline unsigned int bytestream2_put_buffer(PutByteContext *p, const uint8_t *src, unsigned int size)
Definition: bytestream.h:282
the normal 2^n-1 "JPEG" YUV ranges
Definition: pixfmt.h:538
#define AVERROR_PATCHWELCOME
Not yet implemented in FFmpeg, patches welcome.
Definition: error.h:62
static av_always_inline int bytestream2_tell(GetByteContext *g)
Definition: bytestream.h:188
Libavcodec external API header.
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:339
static int init_get_bits8(GetBitContext *s, const uint8_t *buffer, int byte_size)
Initialize GetBitContext.
Definition: get_bits.h:677
main external API structure.
Definition: avcodec.h:526
unsigned a_count_size
Definition: notchlc.c:56
unsigned data_end
Definition: notchlc.c:57
static av_cold int decode_end(AVCodecContext *avctx)
Definition: notchlc.c:519
enum AVColorSpace colorspace
YUV colorspace type.
Definition: avcodec.h:1154
enum AVColorTransferCharacteristic color_trc
Color Transfer Characteristic.
Definition: avcodec.h:1147
#define HISTORY_SIZE
Definition: notchlc.c:74
uint8_t * uncompressed_buffer
Definition: notchlc.c:39
unsigned compressed_size
Definition: notchlc.c:36
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:322
int
IEC 61966-2-1 (sRGB or sYCC)
Definition: pixfmt.h:497
common internal api header.
static int lz4_decompress(AVCodecContext *avctx, GetByteContext *gb, PutByteContext *pb)
Definition: notchlc.c:76
#define bit(string, value)
Definition: cbs_mpeg2.c:58
unsigned uv_offset_data_offset
Definition: notchlc.c:48
#define MKBETAG(a, b, c, d)
Definition: common.h:407
void * priv_data
Definition: avcodec.h:553
unsigned y_data_offset
Definition: notchlc.c:51
int key_frame
1 -> keyframe, 0-> not
Definition: frame.h:386
static av_always_inline int bytestream2_seek(GetByteContext *g, int offset, int whence)
Definition: bytestream.h:208
unsigned texture_size_y
Definition: notchlc.c:46
static float add(float src0, float src1)
#define av_freep(p)
unsigned texture_size_x
Definition: notchlc.c:45
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
This structure stores compressed data.
Definition: packet.h:340
Definition: rpzaenc.c:58
unsigned a_data_offset
Definition: notchlc.c:54
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() for allocating buffers and supports custom allocators.
Definition: codec.h:50
int i
Definition: input.c:407