FFmpeg
cfhd.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2015-2016 Kieran Kunhya <kieran@kunhya.com>
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 /**
22  * @file
23  * Cineform HD video decoder
24  */
25 
26 #include "libavutil/attributes.h"
27 #include "libavutil/buffer.h"
28 #include "libavutil/common.h"
29 #include "libavutil/intreadwrite.h"
30 #include "libavutil/pixdesc.h"
31 
32 #include "avcodec.h"
33 #include "bytestream.h"
34 #include "codec_internal.h"
35 #include "decode.h"
36 #include "get_bits.h"
37 #include "internal.h"
38 #include "thread.h"
39 #include "cfhd.h"
40 
41 #define ALPHA_COMPAND_DC_OFFSET 256
42 #define ALPHA_COMPAND_GAIN 9400
43 
44 static av_cold int cfhd_init(AVCodecContext *avctx)
45 {
46  CFHDContext *s = avctx->priv_data;
47 
48  s->avctx = avctx;
49 
50  for (int i = 0; i < 64; i++) {
51  int val = i;
52 
53  if (val >= 40) {
54  if (val >= 54) {
55  val -= 54;
56  val <<= 2;
57  val += 54;
58  }
59 
60  val -= 40;
61  val <<= 2;
62  val += 40;
63  }
64 
65  s->lut[0][i] = val;
66  }
67 
68  for (int i = 0; i < 256; i++)
69  s->lut[1][i] = i + ((768LL * i * i * i) / (256 * 256 * 256));
70 
71  return ff_cfhd_init_vlcs(s);
72 }
73 
75 {
76  s->subband_num = 0;
77  s->level = 0;
78  s->subband_num_actual = 0;
79 }
80 
82 {
83  s->peak.level = 0;
84  s->peak.offset = 0;
85  memset(&s->peak.base, 0, sizeof(s->peak.base));
86 }
87 
89 {
90  s->coded_width = 0;
91  s->coded_height = 0;
92  s->coded_format = AV_PIX_FMT_YUV422P10;
93  s->cropped_height = 0;
94  s->bpc = 10;
95  s->channel_cnt = 3;
96  s->subband_cnt = SUBBAND_COUNT;
97  s->channel_num = 0;
98  s->lowpass_precision = 16;
99  s->quantisation = 1;
100  s->codebook = 0;
101  s->difference_coding = 0;
102  s->frame_type = 0;
103  s->sample_type = 0;
104  if (s->transform_type != 2)
105  s->transform_type = -1;
108 }
109 
110 static inline int dequant_and_decompand(CFHDContext *s, int level, int quantisation, int codebook)
111 {
112  if (codebook == 0 || codebook == 1) {
113  return s->lut[codebook][abs(level)] * FFSIGN(level) * quantisation;
114  } else
115  return level * quantisation;
116 }
117 
118 static inline void difference_coding(int16_t *band, int width, int height)
119 {
120 
121  int i,j;
122  for (i = 0; i < height; i++) {
123  for (j = 1; j < width; j++) {
124  band[j] += band[j-1];
125  }
126  band += width;
127  }
128 }
129 
130 static inline void peak_table(int16_t *band, Peak *peak, int length)
131 {
132  int i;
133  for (i = 0; i < length; i++)
134  if (abs(band[i]) > peak->level)
135  band[i] = bytestream2_get_le16(&peak->base);
136 }
137 
138 static inline void process_alpha(int16_t *alpha, int width)
139 {
140  int i, channel;
141  for (i = 0; i < width; i++) {
142  channel = alpha[i];
144  channel <<= 3;
146  channel >>= 16;
148  alpha[i] = channel;
149  }
150 }
151 
152 static inline void process_bayer(AVFrame *frame, int bpc)
153 {
154  const int linesize = frame->linesize[0];
155  uint16_t *r = (uint16_t *)frame->data[0];
156  uint16_t *g1 = (uint16_t *)(frame->data[0] + 2);
157  uint16_t *g2 = (uint16_t *)(frame->data[0] + frame->linesize[0]);
158  uint16_t *b = (uint16_t *)(frame->data[0] + frame->linesize[0] + 2);
159  const int mid = 1 << (bpc - 1);
160  const int factor = 1 << (16 - bpc);
161 
162  for (int y = 0; y < frame->height >> 1; y++) {
163  for (int x = 0; x < frame->width; x += 2) {
164  int R, G1, G2, B;
165  int g, rg, bg, gd;
166 
167  g = r[x];
168  rg = g1[x];
169  bg = g2[x];
170  gd = b[x];
171  gd -= mid;
172 
173  R = (rg - mid) * 2 + g;
174  G1 = g + gd;
175  G2 = g - gd;
176  B = (bg - mid) * 2 + g;
177 
178  R = av_clip_uintp2(R * factor, 16);
179  G1 = av_clip_uintp2(G1 * factor, 16);
180  G2 = av_clip_uintp2(G2 * factor, 16);
181  B = av_clip_uintp2(B * factor, 16);
182 
183  r[x] = R;
184  g1[x] = G1;
185  g2[x] = G2;
186  b[x] = B;
187  }
188 
189  r += linesize;
190  g1 += linesize;
191  g2 += linesize;
192  b += linesize;
193  }
194 }
195 
196 static inline void interlaced_vertical_filter(int16_t *output, int16_t *low, int16_t *high,
197  int width, int linesize, int plane)
198 {
199  int i;
200  int16_t even, odd;
201  for (i = 0; i < width; i++) {
202  even = (low[i] - high[i])/2;
203  odd = (low[i] + high[i])/2;
204  output[i] = av_clip_uintp2(even, 10);
205  output[i + linesize] = av_clip_uintp2(odd, 10);
206  }
207 }
208 
209 static inline void inverse_temporal_filter(int16_t *low, int16_t *high, int width)
210 {
211  for (int i = 0; i < width; i++) {
212  int even = (low[i] - high[i]) / 2;
213  int odd = (low[i] + high[i]) / 2;
214 
215  low[i] = even;
216  high[i] = odd;
217  }
218 }
219 
221 {
222  int i, j;
223 
224  for (i = 0; i < FF_ARRAY_ELEMS(s->plane); i++) {
225  Plane *p = &s->plane[i];
226  av_freep(&s->plane[i].idwt_buf);
227  av_freep(&s->plane[i].idwt_tmp);
228  s->plane[i].idwt_size = 0;
229 
230  for (j = 0; j < SUBBAND_COUNT_3D; j++)
231  s->plane[i].subband[j] = NULL;
232 
233  for (j = 0; j < 10; j++)
234  s->plane[i].l_h[j] = NULL;
235 
236  for (j = 0; j < DWT_LEVELS_3D; j++)
237  p->band[j][0].read_ok =
238  p->band[j][1].read_ok =
239  p->band[j][2].read_ok =
240  p->band[j][3].read_ok = 0;
241  }
242  s->a_height = 0;
243  s->a_width = 0;
244  s->a_transform_type = INT_MIN;
245 }
246 
247 static int alloc_buffers(AVCodecContext *avctx)
248 {
249  CFHDContext *s = avctx->priv_data;
250  int i, j, ret, planes, bayer = 0;
251  int chroma_x_shift, chroma_y_shift;
252  unsigned k;
253 
254  if ((ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height)) < 0)
255  return ret;
256  avctx->pix_fmt = s->coded_format;
257 
258  ff_cfhddsp_init(&s->dsp, s->bpc, avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16);
259 
260  if ((ret = av_pix_fmt_get_chroma_sub_sample(s->coded_format,
261  &chroma_x_shift,
262  &chroma_y_shift)) < 0)
263  return ret;
264  planes = av_pix_fmt_count_planes(s->coded_format);
265  if (s->coded_format == AV_PIX_FMT_BAYER_RGGB16) {
266  planes = 4;
267  chroma_x_shift = 1;
268  chroma_y_shift = 1;
269  bayer = 1;
270  }
271 
272  for (i = 0; i < planes; i++) {
273  int w8, h8, w4, h4, w2, h2;
274  int width = (i || bayer) ? s->coded_width >> chroma_x_shift : s->coded_width;
275  int height = (i || bayer) ? s->coded_height >> chroma_y_shift : s->coded_height;
276  ptrdiff_t stride = (FFALIGN(width / 8, 8) + 64) * 8;
277 
278  if (chroma_y_shift && !bayer)
279  height = FFALIGN(height / 8, 2) * 8;
280  s->plane[i].width = width;
281  s->plane[i].height = height;
282  s->plane[i].stride = stride;
283 
284  w8 = FFALIGN(s->plane[i].width / 8, 8) + 64;
285  h8 = FFALIGN(height, 8) / 8;
286  w4 = w8 * 2;
287  h4 = h8 * 2;
288  w2 = w4 * 2;
289  h2 = h4 * 2;
290 
291  if (s->transform_type == 0) {
292  s->plane[i].idwt_size = FFALIGN(height, 8) * stride;
293  s->plane[i].idwt_buf =
294  av_calloc(s->plane[i].idwt_size, sizeof(*s->plane[i].idwt_buf));
295  s->plane[i].idwt_tmp =
296  av_malloc_array(s->plane[i].idwt_size, sizeof(*s->plane[i].idwt_tmp));
297  } else {
298  s->plane[i].idwt_size = FFALIGN(height, 8) * stride * 2;
299  s->plane[i].idwt_buf =
300  av_calloc(s->plane[i].idwt_size, sizeof(*s->plane[i].idwt_buf));
301  s->plane[i].idwt_tmp =
302  av_malloc_array(s->plane[i].idwt_size, sizeof(*s->plane[i].idwt_tmp));
303  }
304 
305  if (!s->plane[i].idwt_buf || !s->plane[i].idwt_tmp)
306  return AVERROR(ENOMEM);
307 
308  s->plane[i].subband[0] = s->plane[i].idwt_buf;
309  s->plane[i].subband[1] = s->plane[i].idwt_buf + 2 * w8 * h8;
310  s->plane[i].subband[2] = s->plane[i].idwt_buf + 1 * w8 * h8;
311  s->plane[i].subband[3] = s->plane[i].idwt_buf + 3 * w8 * h8;
312  s->plane[i].subband[4] = s->plane[i].idwt_buf + 2 * w4 * h4;
313  s->plane[i].subband[5] = s->plane[i].idwt_buf + 1 * w4 * h4;
314  s->plane[i].subband[6] = s->plane[i].idwt_buf + 3 * w4 * h4;
315  if (s->transform_type == 0) {
316  s->plane[i].subband[7] = s->plane[i].idwt_buf + 2 * w2 * h2;
317  s->plane[i].subband[8] = s->plane[i].idwt_buf + 1 * w2 * h2;
318  s->plane[i].subband[9] = s->plane[i].idwt_buf + 3 * w2 * h2;
319  } else {
320  int16_t *frame2 =
321  s->plane[i].subband[7] = s->plane[i].idwt_buf + 4 * w2 * h2;
322  s->plane[i].subband[8] = frame2 + 2 * w4 * h4;
323  s->plane[i].subband[9] = frame2 + 1 * w4 * h4;
324  s->plane[i].subband[10] = frame2 + 3 * w4 * h4;
325  s->plane[i].subband[11] = frame2 + 2 * w2 * h2;
326  s->plane[i].subband[12] = frame2 + 1 * w2 * h2;
327  s->plane[i].subband[13] = frame2 + 3 * w2 * h2;
328  s->plane[i].subband[14] = s->plane[i].idwt_buf + 2 * w2 * h2;
329  s->plane[i].subband[15] = s->plane[i].idwt_buf + 1 * w2 * h2;
330  s->plane[i].subband[16] = s->plane[i].idwt_buf + 3 * w2 * h2;
331  }
332 
333  if (s->transform_type == 0) {
334  for (j = 0; j < DWT_LEVELS; j++) {
335  for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++) {
336  s->plane[i].band[j][k].a_width = w8 << j;
337  s->plane[i].band[j][k].a_height = h8 << j;
338  }
339  }
340  } else {
341  for (j = 0; j < DWT_LEVELS_3D; j++) {
342  int t = j < 1 ? 0 : (j < 3 ? 1 : 2);
343 
344  for (k = 0; k < FF_ARRAY_ELEMS(s->plane[i].band[j]); k++) {
345  s->plane[i].band[j][k].a_width = w8 << t;
346  s->plane[i].band[j][k].a_height = h8 << t;
347  }
348  }
349  }
350 
351  /* ll2 and ll1 commented out because they are done in-place */
352  s->plane[i].l_h[0] = s->plane[i].idwt_tmp;
353  s->plane[i].l_h[1] = s->plane[i].idwt_tmp + 2 * w8 * h8;
354  // s->plane[i].l_h[2] = ll2;
355  s->plane[i].l_h[3] = s->plane[i].idwt_tmp;
356  s->plane[i].l_h[4] = s->plane[i].idwt_tmp + 2 * w4 * h4;
357  // s->plane[i].l_h[5] = ll1;
358  s->plane[i].l_h[6] = s->plane[i].idwt_tmp;
359  s->plane[i].l_h[7] = s->plane[i].idwt_tmp + 2 * w2 * h2;
360  if (s->transform_type != 0) {
361  int16_t *frame2 = s->plane[i].idwt_tmp + 4 * w2 * h2;
362 
363  s->plane[i].l_h[8] = frame2;
364  s->plane[i].l_h[9] = frame2 + 2 * w2 * h2;
365  }
366  }
367 
368  s->a_transform_type = s->transform_type;
369  s->a_height = s->coded_height;
370  s->a_width = s->coded_width;
371  s->a_format = s->coded_format;
372 
373  return 0;
374 }
375 
376 static int cfhd_decode(AVCodecContext *avctx, AVFrame *pic,
377  int *got_frame, AVPacket *avpkt)
378 {
379  CFHDContext *s = avctx->priv_data;
380  CFHDDSPContext *dsp = &s->dsp;
381  GetByteContext gb;
382  int ret = 0, i, j, plane, got_buffer = 0;
383  int16_t *coeff_data;
384 
386  s->planes = av_pix_fmt_count_planes(s->coded_format);
387 
388  bytestream2_init(&gb, avpkt->data, avpkt->size);
389 
390  while (bytestream2_get_bytes_left(&gb) >= 4) {
391  /* Bit weird but implement the tag parsing as the spec says */
392  uint16_t tagu = bytestream2_get_be16(&gb);
393  int16_t tag = (int16_t)tagu;
394  int8_t tag8 = (int8_t)(tagu >> 8);
395  uint16_t abstag = abs(tag);
396  int8_t abs_tag8 = abs(tag8);
397  uint16_t data = bytestream2_get_be16(&gb);
398  if (abs_tag8 >= 0x60 && abs_tag8 <= 0x6f) {
399  av_log(avctx, AV_LOG_DEBUG, "large len %x\n", ((tagu & 0xff) << 16) | data);
400  } else if (tag == SampleFlags) {
401  av_log(avctx, AV_LOG_DEBUG, "Progressive? %"PRIu16"\n", data);
402  s->progressive = data & 0x0001;
403  } else if (tag == FrameType) {
404  s->frame_type = data;
405  av_log(avctx, AV_LOG_DEBUG, "Frame type %"PRIu16"\n", data);
406  } else if (abstag == VersionMajor) {
407  av_log(avctx, AV_LOG_DEBUG, "Version major %"PRIu16"\n", data);
408  } else if (abstag == VersionMinor) {
409  av_log(avctx, AV_LOG_DEBUG, "Version minor %"PRIu16"\n", data);
410  } else if (abstag == VersionRevision) {
411  av_log(avctx, AV_LOG_DEBUG, "Version revision %"PRIu16"\n", data);
412  } else if (abstag == VersionEdit) {
413  av_log(avctx, AV_LOG_DEBUG, "Version edit %"PRIu16"\n", data);
414  } else if (abstag == Version) {
415  av_log(avctx, AV_LOG_DEBUG, "Version %"PRIu16"\n", data);
416  } else if (tag == ImageWidth) {
417  av_log(avctx, AV_LOG_DEBUG, "Width %"PRIu16"\n", data);
418  s->coded_width = data;
419  } else if (tag == ImageHeight) {
420  av_log(avctx, AV_LOG_DEBUG, "Height %"PRIu16"\n", data);
421  s->coded_height = data;
422  } else if (tag == ChannelCount) {
423  av_log(avctx, AV_LOG_DEBUG, "Channel Count: %"PRIu16"\n", data);
424  s->channel_cnt = data;
425  if (data > 4) {
426  av_log(avctx, AV_LOG_ERROR, "Channel Count of %"PRIu16" is unsupported\n", data);
428  goto end;
429  }
430  } else if (tag == SubbandCount) {
431  av_log(avctx, AV_LOG_DEBUG, "Subband Count: %"PRIu16"\n", data);
432  if (data != SUBBAND_COUNT && data != SUBBAND_COUNT_3D) {
433  av_log(avctx, AV_LOG_ERROR, "Subband Count of %"PRIu16" is unsupported\n", data);
435  goto end;
436  }
437  } else if (tag == ChannelNumber) {
438  s->channel_num = data;
439  av_log(avctx, AV_LOG_DEBUG, "Channel number %"PRIu16"\n", data);
440  if (s->channel_num >= s->planes) {
441  av_log(avctx, AV_LOG_ERROR, "Invalid channel number\n");
442  ret = AVERROR(EINVAL);
443  goto end;
444  }
446  } else if (tag == SubbandNumber) {
447  if (s->subband_num != 0 && data == 1 && (s->transform_type == 0 || s->transform_type == 2)) // hack
448  s->level++;
449  av_log(avctx, AV_LOG_DEBUG, "Subband number %"PRIu16"\n", data);
450  s->subband_num = data;
451  if ((s->transform_type == 0 && s->level >= DWT_LEVELS) ||
452  (s->transform_type == 2 && s->level >= DWT_LEVELS_3D)) {
453  av_log(avctx, AV_LOG_ERROR, "Invalid level\n");
454  ret = AVERROR(EINVAL);
455  goto end;
456  }
457  if (s->subband_num > 3) {
458  av_log(avctx, AV_LOG_ERROR, "Invalid subband number\n");
459  ret = AVERROR(EINVAL);
460  goto end;
461  }
462  } else if (tag == SubbandBand) {
463  av_log(avctx, AV_LOG_DEBUG, "Subband number actual %"PRIu16"\n", data);
464  if ((s->transform_type == 0 && data >= SUBBAND_COUNT) ||
465  (s->transform_type == 2 && data >= SUBBAND_COUNT_3D && data != 255)) {
466  av_log(avctx, AV_LOG_ERROR, "Invalid subband number actual\n");
467  ret = AVERROR(EINVAL);
468  goto end;
469  }
470  if (s->transform_type == 0 || s->transform_type == 2)
471  s->subband_num_actual = data;
472  else
473  av_log(avctx, AV_LOG_WARNING, "Ignoring subband num actual %"PRIu16"\n", data);
474  } else if (tag == LowpassPrecision)
475  av_log(avctx, AV_LOG_DEBUG, "Lowpass precision bits: %"PRIu16"\n", data);
476  else if (tag == Quantization) {
477  s->quantisation = data;
478  av_log(avctx, AV_LOG_DEBUG, "Quantisation: %"PRIu16"\n", data);
479  } else if (tag == PrescaleTable) {
480  for (i = 0; i < 8; i++)
481  s->prescale_table[i] = (data >> (14 - i * 2)) & 0x3;
482  av_log(avctx, AV_LOG_DEBUG, "Prescale table: %x\n", data);
483  } else if (tag == BandEncoding) {
484  if (!data || data > 5) {
485  av_log(avctx, AV_LOG_ERROR, "Invalid band encoding\n");
486  ret = AVERROR(EINVAL);
487  goto end;
488  }
489  s->band_encoding = data;
490  av_log(avctx, AV_LOG_DEBUG, "Encode Method for Subband %d : %x\n", s->subband_num_actual, data);
491  } else if (tag == LowpassWidth) {
492  av_log(avctx, AV_LOG_DEBUG, "Lowpass width %"PRIu16"\n", data);
493  s->plane[s->channel_num].band[0][0].width = data;
494  s->plane[s->channel_num].band[0][0].stride = data;
495  } else if (tag == LowpassHeight) {
496  av_log(avctx, AV_LOG_DEBUG, "Lowpass height %"PRIu16"\n", data);
497  s->plane[s->channel_num].band[0][0].height = data;
498  } else if (tag == SampleType) {
499  s->sample_type = data;
500  av_log(avctx, AV_LOG_DEBUG, "Sample type? %"PRIu16"\n", data);
501  } else if (tag == TransformType) {
502  if (data > 2) {
503  av_log(avctx, AV_LOG_ERROR, "Invalid transform type\n");
504  ret = AVERROR(EINVAL);
505  goto end;
506  } else if (data == 1) {
507  av_log(avctx, AV_LOG_ERROR, "unsupported transform type\n");
509  goto end;
510  }
511  if (s->transform_type == -1) {
512  s->transform_type = data;
513  av_log(avctx, AV_LOG_DEBUG, "Transform type %"PRIu16"\n", data);
514  } else {
515  av_log(avctx, AV_LOG_DEBUG, "Ignoring additional transform type %"PRIu16"\n", data);
516  }
517  } else if (abstag >= 0x4000 && abstag <= 0x40ff) {
518  if (abstag == 0x4001)
519  s->peak.level = 0;
520  av_log(avctx, AV_LOG_DEBUG, "Small chunk length %d %s\n", data * 4, tag < 0 ? "optional" : "required");
521  bytestream2_skipu(&gb, data * 4);
522  } else if (tag == FrameIndex) {
523  av_log(avctx, AV_LOG_DEBUG, "Frame index %"PRIu16"\n", data);
524  s->frame_index = data;
525  } else if (tag == SampleIndexTable) {
526  av_log(avctx, AV_LOG_DEBUG, "Sample index table - skipping %i values\n", data);
527  if (data > bytestream2_get_bytes_left(&gb) / 4) {
528  av_log(avctx, AV_LOG_ERROR, "too many values (%d)\n", data);
530  goto end;
531  }
532  for (i = 0; i < data; i++) {
533  uint32_t offset = bytestream2_get_be32(&gb);
534  av_log(avctx, AV_LOG_DEBUG, "Offset = %"PRIu32"\n", offset);
535  }
536  } else if (tag == HighpassWidth) {
537  av_log(avctx, AV_LOG_DEBUG, "Highpass width %i channel %i level %i subband %i\n", data, s->channel_num, s->level, s->subband_num);
538  if (data < 3) {
539  av_log(avctx, AV_LOG_ERROR, "Invalid highpass width\n");
540  ret = AVERROR(EINVAL);
541  goto end;
542  }
543  s->plane[s->channel_num].band[s->level][s->subband_num].width = data;
544  s->plane[s->channel_num].band[s->level][s->subband_num].stride = FFALIGN(data, 8);
545  } else if (tag == HighpassHeight) {
546  av_log(avctx, AV_LOG_DEBUG, "Highpass height %i\n", data);
547  if (data < 3) {
548  av_log(avctx, AV_LOG_ERROR, "Invalid highpass height\n");
549  ret = AVERROR(EINVAL);
550  goto end;
551  }
552  s->plane[s->channel_num].band[s->level][s->subband_num].height = data;
553  } else if (tag == BandWidth) {
554  av_log(avctx, AV_LOG_DEBUG, "Highpass width2 %i\n", data);
555  if (data < 3) {
556  av_log(avctx, AV_LOG_ERROR, "Invalid highpass width2\n");
557  ret = AVERROR(EINVAL);
558  goto end;
559  }
560  s->plane[s->channel_num].band[s->level][s->subband_num].width = data;
561  s->plane[s->channel_num].band[s->level][s->subband_num].stride = FFALIGN(data, 8);
562  } else if (tag == BandHeight) {
563  av_log(avctx, AV_LOG_DEBUG, "Highpass height2 %i\n", data);
564  if (data < 3) {
565  av_log(avctx, AV_LOG_ERROR, "Invalid highpass height2\n");
566  ret = AVERROR(EINVAL);
567  goto end;
568  }
569  s->plane[s->channel_num].band[s->level][s->subband_num].height = data;
570  } else if (tag == InputFormat) {
571  av_log(avctx, AV_LOG_DEBUG, "Input format %i\n", data);
572  if (s->coded_format == AV_PIX_FMT_NONE ||
573  s->coded_format == AV_PIX_FMT_YUV422P10) {
574  if (data >= 100 && data <= 105) {
575  s->coded_format = AV_PIX_FMT_BAYER_RGGB16;
576  } else if (data >= 122 && data <= 128) {
577  s->coded_format = AV_PIX_FMT_GBRP12;
578  } else if (data == 30) {
579  s->coded_format = AV_PIX_FMT_GBRAP12;
580  } else {
581  s->coded_format = AV_PIX_FMT_YUV422P10;
582  }
583  s->planes = s->coded_format == AV_PIX_FMT_BAYER_RGGB16 ? 4 : av_pix_fmt_count_planes(s->coded_format);
584  }
585  } else if (tag == BandCodingFlags) {
586  s->codebook = data & 0xf;
587  s->difference_coding = (data >> 4) & 1;
588  av_log(avctx, AV_LOG_DEBUG, "Other codebook? %i\n", s->codebook);
589  } else if (tag == Precision) {
590  av_log(avctx, AV_LOG_DEBUG, "Precision %i\n", data);
591  if (!(data == 10 || data == 12)) {
592  av_log(avctx, AV_LOG_ERROR, "Invalid bits per channel\n");
593  ret = AVERROR(EINVAL);
594  goto end;
595  }
596  avctx->bits_per_raw_sample = s->bpc = data;
597  } else if (tag == EncodedFormat) {
598  av_log(avctx, AV_LOG_DEBUG, "Sample format? %i\n", data);
599  if (data == 1) {
600  s->coded_format = AV_PIX_FMT_YUV422P10;
601  } else if (data == 2) {
602  s->coded_format = AV_PIX_FMT_BAYER_RGGB16;
603  } else if (data == 3) {
604  s->coded_format = AV_PIX_FMT_GBRP12;
605  } else if (data == 4) {
606  s->coded_format = AV_PIX_FMT_GBRAP12;
607  } else {
608  avpriv_report_missing_feature(avctx, "Sample format of %"PRIu16, data);
610  goto end;
611  }
612  s->planes = data == 2 ? 4 : av_pix_fmt_count_planes(s->coded_format);
613  } else if (tag == -DisplayHeight) {
614  av_log(avctx, AV_LOG_DEBUG, "Cropped height %"PRIu16"\n", data);
615  s->cropped_height = data;
616  } else if (tag == -PeakOffsetLow) {
617  s->peak.offset &= ~0xffff;
618  s->peak.offset |= (data & 0xffff);
619  s->peak.base = gb;
620  s->peak.level = 0;
621  } else if (tag == -PeakOffsetHigh) {
622  s->peak.offset &= 0xffff;
623  s->peak.offset |= (data & 0xffffU)<<16;
624  s->peak.base = gb;
625  s->peak.level = 0;
626  } else if (tag == -PeakLevel && s->peak.offset) {
627  s->peak.level = data;
628  if (s->peak.offset < 4 - bytestream2_tell(&s->peak.base) ||
629  s->peak.offset > 4 + bytestream2_get_bytes_left(&s->peak.base)
630  ) {
632  goto end;
633  }
634  bytestream2_seek(&s->peak.base, s->peak.offset - 4, SEEK_CUR);
635  } else
636  av_log(avctx, AV_LOG_DEBUG, "Unknown tag %i data %x\n", tag, data);
637 
638  if (tag == BitstreamMarker && data == 0xf0f &&
639  s->coded_format != AV_PIX_FMT_NONE) {
640  int lowpass_height = s->plane[s->channel_num].band[0][0].height;
641  int lowpass_width = s->plane[s->channel_num].band[0][0].width;
642  int factor = s->coded_format == AV_PIX_FMT_BAYER_RGGB16 ? 2 : 1;
643 
644  if (s->coded_width) {
645  s->coded_width *= factor;
646  }
647 
648  if (s->coded_height) {
649  s->coded_height *= factor;
650  }
651 
652  if (!s->a_width && !s->coded_width) {
653  s->coded_width = lowpass_width * factor * 8;
654  }
655 
656  if (!s->a_height && !s->coded_height) {
657  s->coded_height = lowpass_height * factor * 8;
658  }
659 
660  if (s->a_width && !s->coded_width)
661  s->coded_width = s->a_width;
662  if (s->a_height && !s->coded_height)
663  s->coded_height = s->a_height;
664 
665  if (s->a_width != s->coded_width || s->a_height != s->coded_height ||
666  s->a_format != s->coded_format ||
667  s->transform_type != s->a_transform_type) {
668  free_buffers(s);
669  if ((ret = alloc_buffers(avctx)) < 0) {
670  free_buffers(s);
671  return ret;
672  }
673  }
674  ret = ff_set_dimensions(avctx, s->coded_width, s->coded_height);
675  if (ret < 0)
676  return ret;
677  if (s->cropped_height) {
678  unsigned height = s->cropped_height << (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16);
679  if (avctx->height < height)
680  return AVERROR_INVALIDDATA;
681  avctx->height = height;
682  }
683  pic->width = pic->height = 0;
684 
685  if ((ret = ff_thread_get_buffer(avctx, pic, 0)) < 0)
686  return ret;
687 
688  s->coded_width = 0;
689  s->coded_height = 0;
690  s->coded_format = AV_PIX_FMT_NONE;
691  got_buffer = 1;
692  } else if (tag == FrameIndex && data == 1 && s->sample_type == 1 && s->frame_type == 2) {
693  pic->width = pic->height = 0;
694 
695  if ((ret = ff_thread_get_buffer(avctx, pic, 0)) < 0)
696  return ret;
697  s->coded_width = 0;
698  s->coded_height = 0;
699  s->coded_format = AV_PIX_FMT_NONE;
700  got_buffer = 1;
701  }
702 
703  if (s->subband_num_actual == 255)
704  goto finish;
705  coeff_data = s->plane[s->channel_num].subband[s->subband_num_actual];
706 
707  /* Lowpass coefficients */
708  if (tag == BitstreamMarker && data == 0xf0f) {
709  int lowpass_height, lowpass_width, lowpass_a_height, lowpass_a_width;
710 
711  if (!s->a_width || !s->a_height) {
713  goto end;
714  }
715 
716  lowpass_height = s->plane[s->channel_num].band[0][0].height;
717  lowpass_width = s->plane[s->channel_num].band[0][0].width;
718  lowpass_a_height = s->plane[s->channel_num].band[0][0].a_height;
719  lowpass_a_width = s->plane[s->channel_num].band[0][0].a_width;
720 
721  if (lowpass_width < 3 ||
722  lowpass_width > lowpass_a_width) {
723  av_log(avctx, AV_LOG_ERROR, "Invalid lowpass width\n");
724  ret = AVERROR(EINVAL);
725  goto end;
726  }
727 
728  if (lowpass_height < 3 ||
729  lowpass_height > lowpass_a_height) {
730  av_log(avctx, AV_LOG_ERROR, "Invalid lowpass height\n");
731  ret = AVERROR(EINVAL);
732  goto end;
733  }
734 
735  if (!got_buffer) {
736  av_log(avctx, AV_LOG_ERROR, "No end of header tag found\n");
737  ret = AVERROR(EINVAL);
738  goto end;
739  }
740 
741  if (lowpass_height > lowpass_a_height || lowpass_width > lowpass_a_width ||
742  lowpass_width * lowpass_height * sizeof(int16_t) > bytestream2_get_bytes_left(&gb)) {
743  av_log(avctx, AV_LOG_ERROR, "Too many lowpass coefficients\n");
744  ret = AVERROR(EINVAL);
745  goto end;
746  }
747 
748  av_log(avctx, AV_LOG_DEBUG, "Start of lowpass coeffs component %d height:%d, width:%d\n", s->channel_num, lowpass_height, lowpass_width);
749  for (i = 0; i < lowpass_height; i++) {
750  for (j = 0; j < lowpass_width; j++)
751  coeff_data[j] = bytestream2_get_be16u(&gb);
752 
753  coeff_data += lowpass_width;
754  }
755 
756  /* Align to mod-4 position to continue reading tags */
757  bytestream2_seek(&gb, bytestream2_tell(&gb) & 3, SEEK_CUR);
758 
759  /* Copy last line of coefficients if odd height */
760  if (lowpass_height & 1) {
761  memcpy(&coeff_data[lowpass_height * lowpass_width],
762  &coeff_data[(lowpass_height - 1) * lowpass_width],
763  lowpass_width * sizeof(*coeff_data));
764  }
765 
766  s->plane[s->channel_num].band[0][0].read_ok = 1;
767 
768  av_log(avctx, AV_LOG_DEBUG, "Lowpass coefficients %d\n", lowpass_width * lowpass_height);
769  }
770 
771  av_assert0(s->subband_num_actual != 255);
772  if (tag == BandHeader || tag == BandSecondPass) {
773  int highpass_height, highpass_width, highpass_a_width, highpass_a_height, highpass_stride, a_expected;
774  int expected;
775  int level, run, coeff;
776  int count = 0, bytes;
777 
778  if (!s->a_width || !s->a_height) {
780  goto end;
781  }
782 
783  highpass_height = s->plane[s->channel_num].band[s->level][s->subband_num].height;
784  highpass_width = s->plane[s->channel_num].band[s->level][s->subband_num].width;
785  highpass_a_width = s->plane[s->channel_num].band[s->level][s->subband_num].a_width;
786  highpass_a_height = s->plane[s->channel_num].band[s->level][s->subband_num].a_height;
787  highpass_stride = s->plane[s->channel_num].band[s->level][s->subband_num].stride;
788  a_expected = highpass_a_height * highpass_a_width;
789 
790  if (!got_buffer) {
791  av_log(avctx, AV_LOG_ERROR, "No end of header tag found\n");
792  ret = AVERROR(EINVAL);
793  goto end;
794  }
795 
796  if (highpass_height > highpass_a_height || highpass_width > highpass_a_width || a_expected < highpass_height * (uint64_t)highpass_stride) {
797  av_log(avctx, AV_LOG_ERROR, "Too many highpass coefficients\n");
798  ret = AVERROR(EINVAL);
799  goto end;
800  }
801  expected = highpass_height * highpass_stride;
802 
803  av_log(avctx, AV_LOG_DEBUG, "Start subband coeffs plane %i level %i codebook %i expected %i\n", s->channel_num, s->level, s->codebook, expected);
804 
806  if (ret < 0)
807  goto end;
808  {
809  OPEN_READER(re, &s->gb);
810 
811  const int lossless = s->band_encoding == 5;
812 
813  if (s->codebook == 0 && s->transform_type == 2 && s->subband_num_actual == 7)
814  s->codebook = 1;
815  if (!s->codebook) {
816  while (1) {
817  UPDATE_CACHE(re, &s->gb);
818  GET_RL_VLC(level, run, re, &s->gb, s->table_9_rl_vlc,
819  VLC_BITS, 3, 1);
820 
821  /* escape */
822  if (!run)
823  break;
824 
825  count += run;
826 
827  if (count > expected)
828  break;
829 
830  if (!lossless)
831  coeff = dequant_and_decompand(s, level, s->quantisation, 0);
832  else
833  coeff = level;
834  if (tag == BandSecondPass) {
835  const uint16_t q = s->quantisation;
836 
837  for (i = 0; i < run; i++) {
838  *coeff_data |= coeff * 256U;
839  *coeff_data++ *= q;
840  }
841  } else {
842  for (i = 0; i < run; i++)
843  *coeff_data++ = coeff;
844  }
845  }
846  } else {
847  while (1) {
848  UPDATE_CACHE(re, &s->gb);
849  GET_RL_VLC(level, run, re, &s->gb, s->table_18_rl_vlc,
850  VLC_BITS, 3, 1);
851 
852  /* escape */
853  if (!run)
854  break;
855 
856  count += run;
857 
858  if (count > expected)
859  break;
860 
861  if (!lossless)
862  coeff = dequant_and_decompand(s, level, s->quantisation, s->codebook);
863  else
864  coeff = level;
865  if (tag == BandSecondPass) {
866  const uint16_t q = s->quantisation;
867 
868  for (i = 0; i < run; i++) {
869  *coeff_data |= coeff * 256U;
870  *coeff_data++ *= q;
871  }
872  } else {
873  for (i = 0; i < run; i++)
874  *coeff_data++ = coeff;
875  }
876  }
877  }
878  CLOSE_READER(re, &s->gb);
879  }
880 
881  if (count > expected) {
882  av_log(avctx, AV_LOG_ERROR, "Escape codeword not found, probably corrupt data\n");
883  ret = AVERROR(EINVAL);
884  goto end;
885  }
886  if (s->peak.level)
887  peak_table(coeff_data - count, &s->peak, count);
888  if (s->difference_coding)
889  difference_coding(s->plane[s->channel_num].subband[s->subband_num_actual], highpass_width, highpass_height);
890 
891  bytes = FFALIGN(AV_CEIL_RSHIFT(get_bits_count(&s->gb), 3), 4);
892  if (bytes > bytestream2_get_bytes_left(&gb)) {
893  av_log(avctx, AV_LOG_ERROR, "Bitstream overread error\n");
894  ret = AVERROR(EINVAL);
895  goto end;
896  } else
897  bytestream2_seek(&gb, bytes, SEEK_CUR);
898 
899  av_log(avctx, AV_LOG_DEBUG, "End subband coeffs %i extra %i\n", count, count - expected);
900  s->plane[s->channel_num].band[s->level][s->subband_num].read_ok = 1;
901 finish:
902  if (s->subband_num_actual != 255)
903  s->codebook = 0;
904  }
905  }
906 
907  s->planes = av_pix_fmt_count_planes(avctx->pix_fmt);
908  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
909  s->progressive = 1;
910  s->planes = 4;
911  }
912 
913  ff_thread_finish_setup(avctx);
914 
915  if (!s->a_width || !s->a_height || s->a_format == AV_PIX_FMT_NONE ||
916  s->a_transform_type == INT_MIN ||
917  s->coded_width || s->coded_height || s->coded_format != AV_PIX_FMT_NONE) {
918  av_log(avctx, AV_LOG_ERROR, "Invalid dimensions\n");
919  ret = AVERROR(EINVAL);
920  goto end;
921  }
922 
923  if (!got_buffer) {
924  av_log(avctx, AV_LOG_ERROR, "No end of header tag found\n");
925  ret = AVERROR(EINVAL);
926  goto end;
927  }
928 
929  for (plane = 0; plane < s->planes; plane++) {
930  int o, level;
931 
932  for (level = 0; level < (s->transform_type == 0 ? DWT_LEVELS : DWT_LEVELS_3D) ; level++) {
933  if (s->transform_type == 2)
934  if (level == 2 || level == 5)
935  continue;
936  for (o = !!level; o < 4 ; o++) {
937  if (!s->plane[plane].band[level][o].read_ok) {
939  goto end;
940  }
941  }
942  }
943  }
944 
945  if (s->transform_type == 0 && s->sample_type != 1) {
946  for (plane = 0; plane < s->planes && !ret; plane++) {
947  /* level 1 */
948  int lowpass_height = s->plane[plane].band[0][0].height;
949  int output_stride = s->plane[plane].band[0][0].a_width;
950  int lowpass_width = s->plane[plane].band[0][0].width;
951  int highpass_stride = s->plane[plane].band[0][1].stride;
952  int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
953  ptrdiff_t dst_linesize;
954  int16_t *low, *high, *output, *dst;
955 
956  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
957  act_plane = 0;
958  dst_linesize = pic->linesize[act_plane];
959  } else {
960  dst_linesize = pic->linesize[act_plane] / 2;
961  }
962 
963  if (lowpass_height > s->plane[plane].band[0][0].a_height || lowpass_width > s->plane[plane].band[0][0].a_width ||
964  !highpass_stride || s->plane[plane].band[0][1].width > s->plane[plane].band[0][1].a_width ||
965  lowpass_width < 3 || lowpass_height < 3) {
966  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
967  ret = AVERROR(EINVAL);
968  goto end;
969  }
970 
971  av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
972 
973  low = s->plane[plane].subband[0];
974  high = s->plane[plane].subband[2];
975  output = s->plane[plane].l_h[0];
976  dsp->vert_filter(output, output_stride, low, lowpass_width, high, highpass_stride, lowpass_width, lowpass_height);
977 
978  low = s->plane[plane].subband[1];
979  high = s->plane[plane].subband[3];
980  output = s->plane[plane].l_h[1];
981 
982  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
983 
984  low = s->plane[plane].l_h[0];
985  high = s->plane[plane].l_h[1];
986  output = s->plane[plane].subband[0];
987  dsp->horiz_filter(output, output_stride, low, output_stride, high, output_stride, lowpass_width, lowpass_height * 2);
988  if (s->bpc == 12) {
989  output = s->plane[plane].subband[0];
990  for (i = 0; i < lowpass_height * 2; i++) {
991  for (j = 0; j < lowpass_width * 2; j++)
992  output[j] *= 4;
993 
994  output += output_stride * 2;
995  }
996  }
997 
998  /* level 2 */
999  lowpass_height = s->plane[plane].band[1][1].height;
1000  output_stride = s->plane[plane].band[1][1].a_width;
1001  lowpass_width = s->plane[plane].band[1][1].width;
1002  highpass_stride = s->plane[plane].band[1][1].stride;
1003 
1004  if (lowpass_height > s->plane[plane].band[1][1].a_height || lowpass_width > s->plane[plane].band[1][1].a_width ||
1005  !highpass_stride || s->plane[plane].band[1][1].width > s->plane[plane].band[1][1].a_width ||
1006  lowpass_width < 3 || lowpass_height < 3) {
1007  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1008  ret = AVERROR(EINVAL);
1009  goto end;
1010  }
1011 
1012  av_log(avctx, AV_LOG_DEBUG, "Level 2 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
1013 
1014  low = s->plane[plane].subband[0];
1015  high = s->plane[plane].subband[5];
1016  output = s->plane[plane].l_h[3];
1017  dsp->vert_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1018 
1019  low = s->plane[plane].subband[4];
1020  high = s->plane[plane].subband[6];
1021  output = s->plane[plane].l_h[4];
1022  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1023 
1024  low = s->plane[plane].l_h[3];
1025  high = s->plane[plane].l_h[4];
1026  output = s->plane[plane].subband[0];
1027  dsp->horiz_filter(output, output_stride, low, output_stride, high, output_stride, lowpass_width, lowpass_height * 2);
1028 
1029  output = s->plane[plane].subband[0];
1030  for (i = 0; i < lowpass_height * 2; i++) {
1031  for (j = 0; j < lowpass_width * 2; j++)
1032  output[j] *= 4;
1033 
1034  output += output_stride * 2;
1035  }
1036 
1037  /* level 3 */
1038  lowpass_height = s->plane[plane].band[2][1].height;
1039  output_stride = s->plane[plane].band[2][1].a_width;
1040  lowpass_width = s->plane[plane].band[2][1].width;
1041  highpass_stride = s->plane[plane].band[2][1].stride;
1042 
1043  if (lowpass_height > s->plane[plane].band[2][1].a_height || lowpass_width > s->plane[plane].band[2][1].a_width ||
1044  !highpass_stride || s->plane[plane].band[2][1].width > s->plane[plane].band[2][1].a_width ||
1045  lowpass_height < 3 || lowpass_width < 3 || lowpass_width * 2 > s->plane[plane].width) {
1046  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1047  ret = AVERROR(EINVAL);
1048  goto end;
1049  }
1050 
1051  av_log(avctx, AV_LOG_DEBUG, "Level 3 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
1052  if (s->progressive) {
1053  low = s->plane[plane].subband[0];
1054  high = s->plane[plane].subband[8];
1055  output = s->plane[plane].l_h[6];
1056  dsp->vert_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1057 
1058  low = s->plane[plane].subband[7];
1059  high = s->plane[plane].subband[9];
1060  output = s->plane[plane].l_h[7];
1061  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1062 
1063  dst = (int16_t *)pic->data[act_plane];
1064  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
1065  if (plane & 1)
1066  dst++;
1067  if (plane > 1)
1068  dst += pic->linesize[act_plane] >> 1;
1069  }
1070  low = s->plane[plane].l_h[6];
1071  high = s->plane[plane].l_h[7];
1072 
1073  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16 &&
1074  (lowpass_height * 2 > avctx->coded_height / 2 ||
1075  lowpass_width * 2 > avctx->coded_width / 2 )
1076  ) {
1078  goto end;
1079  }
1080 
1081  for (i = 0; i < s->plane[act_plane].height; i++) {
1082  dsp->horiz_filter_clip(dst, low, high, lowpass_width, s->bpc);
1083  if (avctx->pix_fmt == AV_PIX_FMT_GBRAP12 && act_plane == 3)
1084  process_alpha(dst, lowpass_width * 2);
1085  low += output_stride;
1086  high += output_stride;
1087  dst += dst_linesize;
1088  }
1089  } else {
1090  av_log(avctx, AV_LOG_DEBUG, "interlaced frame ? %d", pic->interlaced_frame);
1091  pic->interlaced_frame = 1;
1092  low = s->plane[plane].subband[0];
1093  high = s->plane[plane].subband[7];
1094  output = s->plane[plane].l_h[6];
1095  dsp->horiz_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1096 
1097  low = s->plane[plane].subband[8];
1098  high = s->plane[plane].subband[9];
1099  output = s->plane[plane].l_h[7];
1100  dsp->horiz_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1101 
1102  dst = (int16_t *)pic->data[act_plane];
1103  low = s->plane[plane].l_h[6];
1104  high = s->plane[plane].l_h[7];
1105  for (i = 0; i < s->plane[act_plane].height / 2; i++) {
1106  interlaced_vertical_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, act_plane);
1107  low += output_stride * 2;
1108  high += output_stride * 2;
1109  dst += pic->linesize[act_plane];
1110  }
1111  }
1112  }
1113  } else if (s->transform_type == 2 && (avctx->internal->is_copy || s->frame_index == 1 || s->sample_type != 1)) {
1114  for (plane = 0; plane < s->planes && !ret; plane++) {
1115  int lowpass_height = s->plane[plane].band[0][0].height;
1116  int output_stride = s->plane[plane].band[0][0].a_width;
1117  int lowpass_width = s->plane[plane].band[0][0].width;
1118  int highpass_stride = s->plane[plane].band[0][1].stride;
1119  int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
1120  int16_t *low, *high, *output, *dst;
1121  ptrdiff_t dst_linesize;
1122 
1123  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
1124  act_plane = 0;
1125  dst_linesize = pic->linesize[act_plane];
1126  } else {
1127  dst_linesize = pic->linesize[act_plane] / 2;
1128  }
1129 
1130  if (lowpass_height > s->plane[plane].band[0][0].a_height || lowpass_width > s->plane[plane].band[0][0].a_width ||
1131  !highpass_stride || s->plane[plane].band[0][1].width > s->plane[plane].band[0][1].a_width ||
1132  lowpass_width < 3 || lowpass_height < 3) {
1133  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1134  ret = AVERROR(EINVAL);
1135  goto end;
1136  }
1137 
1138  av_log(avctx, AV_LOG_DEBUG, "Decoding level 1 plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
1139 
1140  low = s->plane[plane].subband[0];
1141  high = s->plane[plane].subband[2];
1142  output = s->plane[plane].l_h[0];
1143  dsp->vert_filter(output, output_stride, low, lowpass_width, high, highpass_stride, lowpass_width, lowpass_height);
1144 
1145  low = s->plane[plane].subband[1];
1146  high = s->plane[plane].subband[3];
1147  output = s->plane[plane].l_h[1];
1148  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1149 
1150  low = s->plane[plane].l_h[0];
1151  high = s->plane[plane].l_h[1];
1152  output = s->plane[plane].l_h[7];
1153  dsp->horiz_filter(output, output_stride, low, output_stride, high, output_stride, lowpass_width, lowpass_height * 2);
1154  if (s->bpc == 12) {
1155  output = s->plane[plane].l_h[7];
1156  for (i = 0; i < lowpass_height * 2; i++) {
1157  for (j = 0; j < lowpass_width * 2; j++)
1158  output[j] *= 4;
1159 
1160  output += output_stride * 2;
1161  }
1162  }
1163 
1164  lowpass_height = s->plane[plane].band[1][1].height;
1165  output_stride = s->plane[plane].band[1][1].a_width;
1166  lowpass_width = s->plane[plane].band[1][1].width;
1167  highpass_stride = s->plane[plane].band[1][1].stride;
1168 
1169  if (lowpass_height > s->plane[plane].band[1][1].a_height || lowpass_width > s->plane[plane].band[1][1].a_width ||
1170  !highpass_stride || s->plane[plane].band[1][1].width > s->plane[plane].band[1][1].a_width ||
1171  lowpass_width < 3 || lowpass_height < 3) {
1172  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1173  ret = AVERROR(EINVAL);
1174  goto end;
1175  }
1176 
1177  av_log(avctx, AV_LOG_DEBUG, "Level 2 lowpass plane %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
1178 
1179  low = s->plane[plane].l_h[7];
1180  high = s->plane[plane].subband[5];
1181  output = s->plane[plane].l_h[3];
1182  dsp->vert_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1183 
1184  low = s->plane[plane].subband[4];
1185  high = s->plane[plane].subband[6];
1186  output = s->plane[plane].l_h[4];
1187  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1188 
1189  low = s->plane[plane].l_h[3];
1190  high = s->plane[plane].l_h[4];
1191  output = s->plane[plane].l_h[7];
1192  dsp->horiz_filter(output, output_stride, low, output_stride, high, output_stride, lowpass_width, lowpass_height * 2);
1193 
1194  output = s->plane[plane].l_h[7];
1195  for (i = 0; i < lowpass_height * 2; i++) {
1196  for (j = 0; j < lowpass_width * 2; j++)
1197  output[j] *= 4;
1198  output += output_stride * 2;
1199  }
1200 
1201  low = s->plane[plane].subband[7];
1202  high = s->plane[plane].subband[9];
1203  output = s->plane[plane].l_h[3];
1204  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1205 
1206  low = s->plane[plane].subband[8];
1207  high = s->plane[plane].subband[10];
1208  output = s->plane[plane].l_h[4];
1209  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1210 
1211  low = s->plane[plane].l_h[3];
1212  high = s->plane[plane].l_h[4];
1213  output = s->plane[plane].l_h[9];
1214  dsp->horiz_filter(output, output_stride, low, output_stride, high, output_stride, lowpass_width, lowpass_height * 2);
1215 
1216  lowpass_height = s->plane[plane].band[4][1].height;
1217  output_stride = s->plane[plane].band[4][1].a_width;
1218  lowpass_width = s->plane[plane].band[4][1].width;
1219  highpass_stride = s->plane[plane].band[4][1].stride;
1220  av_log(avctx, AV_LOG_DEBUG, "temporal level %i %i %i %i\n", plane, lowpass_height, lowpass_width, highpass_stride);
1221 
1222  if (lowpass_height > s->plane[plane].band[4][1].a_height || lowpass_width > s->plane[plane].band[4][1].a_width ||
1223  !highpass_stride || s->plane[plane].band[4][1].width > s->plane[plane].band[4][1].a_width ||
1224  lowpass_width < 3 || lowpass_height < 3) {
1225  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1226  ret = AVERROR(EINVAL);
1227  goto end;
1228  }
1229 
1230  low = s->plane[plane].l_h[7];
1231  high = s->plane[plane].l_h[9];
1232  output = s->plane[plane].l_h[7];
1233  for (i = 0; i < lowpass_height; i++) {
1234  inverse_temporal_filter(low, high, lowpass_width);
1235  low += output_stride;
1236  high += output_stride;
1237  }
1238  if (s->progressive) {
1239  low = s->plane[plane].l_h[7];
1240  high = s->plane[plane].subband[15];
1241  output = s->plane[plane].l_h[6];
1242  dsp->vert_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1243 
1244  low = s->plane[plane].subband[14];
1245  high = s->plane[plane].subband[16];
1246  output = s->plane[plane].l_h[7];
1247  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1248 
1249  low = s->plane[plane].l_h[9];
1250  high = s->plane[plane].subband[12];
1251  output = s->plane[plane].l_h[8];
1252  dsp->vert_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1253 
1254  low = s->plane[plane].subband[11];
1255  high = s->plane[plane].subband[13];
1256  output = s->plane[plane].l_h[9];
1257  dsp->vert_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1258 
1259  if (s->sample_type == 1)
1260  continue;
1261 
1262  dst = (int16_t *)pic->data[act_plane];
1263  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
1264  if (plane & 1)
1265  dst++;
1266  if (plane > 1)
1267  dst += pic->linesize[act_plane] >> 1;
1268  }
1269 
1270  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16 &&
1271  (lowpass_height * 2 > avctx->coded_height / 2 ||
1272  lowpass_width * 2 > avctx->coded_width / 2 )
1273  ) {
1275  goto end;
1276  }
1277 
1278  low = s->plane[plane].l_h[6];
1279  high = s->plane[plane].l_h[7];
1280  for (i = 0; i < s->plane[act_plane].height; i++) {
1281  dsp->horiz_filter_clip(dst, low, high, lowpass_width, s->bpc);
1282  low += output_stride;
1283  high += output_stride;
1284  dst += dst_linesize;
1285  }
1286  } else {
1287  pic->interlaced_frame = 1;
1288  low = s->plane[plane].l_h[7];
1289  high = s->plane[plane].subband[14];
1290  output = s->plane[plane].l_h[6];
1291  dsp->horiz_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1292 
1293  low = s->plane[plane].subband[15];
1294  high = s->plane[plane].subband[16];
1295  output = s->plane[plane].l_h[7];
1296  dsp->horiz_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1297 
1298  low = s->plane[plane].l_h[9];
1299  high = s->plane[plane].subband[11];
1300  output = s->plane[plane].l_h[8];
1301  dsp->horiz_filter(output, output_stride, low, output_stride, high, highpass_stride, lowpass_width, lowpass_height);
1302 
1303  low = s->plane[plane].subband[12];
1304  high = s->plane[plane].subband[13];
1305  output = s->plane[plane].l_h[9];
1306  dsp->horiz_filter(output, output_stride, low, highpass_stride, high, highpass_stride, lowpass_width, lowpass_height);
1307 
1308  if (s->sample_type == 1)
1309  continue;
1310 
1311  dst = (int16_t *)pic->data[act_plane];
1312  low = s->plane[plane].l_h[6];
1313  high = s->plane[plane].l_h[7];
1314  for (i = 0; i < s->plane[act_plane].height / 2; i++) {
1315  interlaced_vertical_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, act_plane);
1316  low += output_stride * 2;
1317  high += output_stride * 2;
1318  dst += pic->linesize[act_plane];
1319  }
1320  }
1321  }
1322  }
1323 
1324  if (s->transform_type == 2 && s->sample_type == 1) {
1325  int16_t *low, *high, *dst;
1326  int output_stride, lowpass_height, lowpass_width;
1327  ptrdiff_t dst_linesize;
1328 
1329  for (plane = 0; plane < s->planes; plane++) {
1330  int act_plane = plane == 1 ? 2 : plane == 2 ? 1 : plane;
1331 
1332  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
1333  act_plane = 0;
1334  dst_linesize = pic->linesize[act_plane];
1335  } else {
1336  dst_linesize = pic->linesize[act_plane] / 2;
1337  }
1338 
1339  lowpass_height = s->plane[plane].band[4][1].height;
1340  output_stride = s->plane[plane].band[4][1].a_width;
1341  lowpass_width = s->plane[plane].band[4][1].width;
1342 
1343  if (lowpass_height > s->plane[plane].band[4][1].a_height || lowpass_width > s->plane[plane].band[4][1].a_width ||
1344  s->plane[plane].band[4][1].width > s->plane[plane].band[4][1].a_width ||
1345  lowpass_width < 3 || lowpass_height < 3) {
1346  av_log(avctx, AV_LOG_ERROR, "Invalid plane dimensions\n");
1347  ret = AVERROR(EINVAL);
1348  goto end;
1349  }
1350 
1351  if (s->progressive) {
1352  dst = (int16_t *)pic->data[act_plane];
1353  low = s->plane[plane].l_h[8];
1354  high = s->plane[plane].l_h[9];
1355 
1356  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16) {
1357  if (plane & 1)
1358  dst++;
1359  if (plane > 1)
1360  dst += pic->linesize[act_plane] >> 1;
1361  }
1362 
1363  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16 &&
1364  (lowpass_height * 2 > avctx->coded_height / 2 ||
1365  lowpass_width * 2 > avctx->coded_width / 2 )
1366  ) {
1368  goto end;
1369  }
1370 
1371  for (i = 0; i < s->plane[act_plane].height; i++) {
1372  dsp->horiz_filter_clip(dst, low, high, lowpass_width, s->bpc);
1373  low += output_stride;
1374  high += output_stride;
1375  dst += dst_linesize;
1376  }
1377  } else {
1378  dst = (int16_t *)pic->data[act_plane];
1379  low = s->plane[plane].l_h[8];
1380  high = s->plane[plane].l_h[9];
1381  for (i = 0; i < s->plane[act_plane].height / 2; i++) {
1382  interlaced_vertical_filter(dst, low, high, lowpass_width * 2, pic->linesize[act_plane]/2, act_plane);
1383  low += output_stride * 2;
1384  high += output_stride * 2;
1385  dst += pic->linesize[act_plane];
1386  }
1387  }
1388  }
1389  }
1390 
1391  if (avctx->pix_fmt == AV_PIX_FMT_BAYER_RGGB16)
1392  process_bayer(pic, s->bpc);
1393 end:
1394  if (ret < 0)
1395  return ret;
1396 
1397  *got_frame = 1;
1398  return avpkt->size;
1399 }
1400 
1402 {
1403  CFHDContext *s = avctx->priv_data;
1404 
1405  free_buffers(s);
1406 
1407  return 0;
1408 }
1409 
1410 #if HAVE_THREADS
1411 static int update_thread_context(AVCodecContext *dst, const AVCodecContext *src)
1412 {
1413  CFHDContext *psrc = src->priv_data;
1414  CFHDContext *pdst = dst->priv_data;
1415  int ret;
1416 
1417  if (dst == src || psrc->transform_type == 0)
1418  return 0;
1419 
1420  if (pdst->plane[0].idwt_size != psrc->plane[0].idwt_size ||
1421  pdst->a_format != psrc->a_format ||
1422  pdst->a_width != psrc->a_width ||
1423  pdst->a_height != psrc->a_height ||
1424  pdst->a_transform_type != psrc->a_transform_type)
1425  free_buffers(pdst);
1426 
1427  pdst->a_format = psrc->a_format;
1428  pdst->a_width = psrc->a_width;
1429  pdst->a_height = psrc->a_height;
1430  pdst->a_transform_type = psrc->a_transform_type;
1431  pdst->transform_type = psrc->transform_type;
1432  pdst->progressive = psrc->progressive;
1433  pdst->planes = psrc->planes;
1434 
1435  if (!pdst->plane[0].idwt_buf) {
1436  pdst->coded_width = pdst->a_width;
1437  pdst->coded_height = pdst->a_height;
1438  pdst->coded_format = pdst->a_format;
1439  pdst->transform_type = pdst->a_transform_type;
1440  ret = alloc_buffers(dst);
1441  if (ret < 0)
1442  return ret;
1443  }
1444 
1445  for (int plane = 0; plane < pdst->planes; plane++) {
1446  memcpy(pdst->plane[plane].band, psrc->plane[plane].band, sizeof(pdst->plane[plane].band));
1447  memcpy(pdst->plane[plane].idwt_buf, psrc->plane[plane].idwt_buf,
1448  pdst->plane[plane].idwt_size * sizeof(int16_t));
1449  }
1450 
1451  return 0;
1452 }
1453 #endif
1454 
1456  .p.name = "cfhd",
1457  CODEC_LONG_NAME("GoPro CineForm HD"),
1458  .p.type = AVMEDIA_TYPE_VIDEO,
1459  .p.id = AV_CODEC_ID_CFHD,
1460  .priv_data_size = sizeof(CFHDContext),
1461  .init = cfhd_init,
1462  .close = cfhd_close,
1465  .p.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_FRAME_THREADS,
1466  .caps_internal = FF_CODEC_CAP_INIT_CLEANUP,
1467 };
ChannelNumber
@ ChannelNumber
Definition: cfhd.h:77
ChannelCount
@ ChannelCount
Definition: cfhd.h:41
AV_LOG_WARNING
#define AV_LOG_WARNING
Something somehow does not look correct.
Definition: log.h:186
BandSecondPass
@ BandSecondPass
Definition: cfhd.h:87
level
uint8_t level
Definition: svq3.c:204
Precision
@ Precision
Definition: cfhd.h:80
FF_CODEC_CAP_INIT_CLEANUP
#define FF_CODEC_CAP_INIT_CLEANUP
The codec allows calling the close function for deallocation even if the init function returned a fai...
Definition: codec_internal.h:42
Peak::level
int level
Definition: cfhd.h:134
r
const char * r
Definition: vf_curves.c:126
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
BandHeader
@ BandHeader
Definition: cfhd.h:75
PrescaleTable
@ PrescaleTable
Definition: cfhd.h:88
GetByteContext
Definition: bytestream.h:33
CFHDContext::progressive
int progressive
Definition: cfhd.h:158
BandHeight
@ BandHeight
Definition: cfhd.h:70
ff_cfhd_decoder
const FFCodec ff_cfhd_decoder
Definition: cfhd.c:1455
av_clip_uintp2
#define av_clip_uintp2
Definition: common.h:119
bytestream2_skipu
static av_always_inline void bytestream2_skipu(GetByteContext *g, unsigned int size)
Definition: bytestream.h:174
output
filter_frame For filters that do not use the this method is called when a frame is pushed to the filter s input It can be called at any time except in a reentrant way If the input frame is enough to produce output
Definition: filter_design.txt:225
get_bits_count
static int get_bits_count(const GetBitContext *s)
Definition: get_bits.h:219
SampleType
int32_t SampleType
Definition: ac3enc.h:67
bytestream2_seek
static av_always_inline int bytestream2_seek(GetByteContext *g, int offset, int whence)
Definition: bytestream.h:212
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:325
pixdesc.h
AVFrame::width
int width
Definition: frame.h:397
CFHDDSPContext
Definition: cfhddsp.h:25
CFHDDSPContext::horiz_filter_clip
void(* horiz_filter_clip)(int16_t *output, const int16_t *low, const int16_t *high, int width, int bpc)
Definition: cfhddsp.h:36
internal.h
even
Tag MUST be even
Definition: snow.txt:206
AVPacket::data
uint8_t * data
Definition: packet.h:374
CFHDContext::a_format
int a_format
Definition: cfhd.h:162
b
#define b
Definition: input.c:41
HighpassWidth
@ HighpassWidth
Definition: cfhd.h:62
data
const char data[16]
Definition: mxf.c:146
R
#define R
Definition: huffyuv.h:44
FFCodec
Definition: codec_internal.h:119
ALPHA_COMPAND_DC_OFFSET
#define ALPHA_COMPAND_DC_OFFSET
Definition: cfhd.c:41
UPDATE_CACHE
#define UPDATE_CACHE(name, gb)
Definition: get_bits.h:178
LowpassWidth
@ LowpassWidth
Definition: cfhd.h:53
ff_set_dimensions
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
Definition: utils.c:91
cfhd_init
static av_cold int cfhd_init(AVCodecContext *avctx)
Definition: cfhd.c:44
difference_coding
static void difference_coding(int16_t *band, int width, int height)
Definition: cfhd.c:118
thread.h
AVFrame::data
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:346
ff_cfhd_init_vlcs
int ff_cfhd_init_vlcs(CFHDContext *s)
Definition: cfhddata.c:181
AVCodecInternal::is_copy
int is_copy
When using frame-threaded decoding, this field is set for the first worker thread (e....
Definition: internal.h:57
av_pix_fmt_count_planes
int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2898
VersionMajor
@ VersionMajor
Definition: cfhd.h:35
FFCodec::p
AVCodec p
The public AVCodec.
Definition: codec_internal.h:123
finish
static void finish(void)
Definition: movenc.c:342
init_peak_table_defaults
static void init_peak_table_defaults(CFHDContext *s)
Definition: cfhd.c:81
cfhd.h
FFSIGN
#define FFSIGN(a)
Definition: common.h:65
ff_thread_get_buffer
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before as well as code calling up to before the decode process starts Call have so the codec calls ff_thread_report set FF_CODEC_CAP_ALLOCATE_PROGRESS in AVCodec caps_internal and use ff_thread_get_buffer() to allocate frames. The frames must then be freed with ff_thread_release_buffer(). Otherwise decode directly into the user-supplied frames. Call ff_thread_report_progress() after some part of the current picture has decoded. A good place to put this is where draw_horiz_band() is called - add this if it isn 't called anywhere
DWT_LEVELS
#define DWT_LEVELS
Definition: cfhd.h:105
val
static double val(void *priv, double ch)
Definition: aeval.c:77
av_pix_fmt_get_chroma_sub_sample
int av_pix_fmt_get_chroma_sub_sample(enum AVPixelFormat pix_fmt, int *h_shift, int *v_shift)
Utility function to access log2_chroma_w log2_chroma_h from the pixel format AVPixFmtDescriptor.
Definition: pixdesc.c:2886
AVCodecContext::coded_height
int coded_height
Definition: avcodec.h:586
LowpassPrecision
@ LowpassPrecision
Definition: cfhd.h:57
Quantization
@ Quantization
Definition: cfhd.h:73
BandEncoding
@ BandEncoding
Definition: cfhd.h:72
SubbandNumber
@ SubbandNumber
Definition: cfhd.h:68
AV_LOG_ERROR
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
Definition: log.h:180
FF_ARRAY_ELEMS
#define FF_ARRAY_ELEMS(a)
Definition: sinewin_tablegen.c:29
av_cold
#define av_cold
Definition: attributes.h:90
peak_table
static void peak_table(int16_t *band, Peak *peak, int length)
Definition: cfhd.c:130
init_get_bits8
static int init_get_bits8(GetBitContext *s, const uint8_t *buffer, int byte_size)
Initialize GetBitContext.
Definition: get_bits.h:667
Plane::idwt_size
int idwt_size
Definition: cfhd.h:124
process_alpha
static void process_alpha(int16_t *alpha, int width)
Definition: cfhd.c:138
AV_CODEC_ID_CFHD
@ AV_CODEC_ID_CFHD
Definition: codec_id.h:270
CLOSE_READER
#define CLOSE_READER(name, gb)
Definition: get_bits.h:149
width
#define width
FF_CODEC_DECODE_CB
#define FF_CODEC_DECODE_CB(func)
Definition: codec_internal.h:298
intreadwrite.h
Version
@ Version
Definition: cfhd.h:86
s
#define s(width, name)
Definition: cbs_vp9.c:256
AV_PIX_FMT_GBRAP12
#define AV_PIX_FMT_GBRAP12
Definition: pixfmt.h:480
ALPHA_COMPAND_GAIN
#define ALPHA_COMPAND_GAIN
Definition: cfhd.c:42
TransformType
TransformType
Definition: webp.c:112
AV_CEIL_RSHIFT
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:50
g
const char * g
Definition: vf_curves.c:127
PeakOffsetHigh
@ PeakOffsetHigh
Definition: cfhd.h:85
GetByteContext::buffer
const uint8_t * buffer
Definition: bytestream.h:34
VersionMinor
@ VersionMinor
Definition: cfhd.h:36
init
int(* init)(AVBSFContext *ctx)
Definition: dts2pts_bsf.c:363
av_assert0
#define av_assert0(cond)
assert() equivalent, that is always enabled.
Definition: avassert.h:37
B
#define B
Definition: huffyuv.h:42
AVCodecContext::bits_per_raw_sample
int bits_per_raw_sample
Bits per sample/pixel of internal libavcodec pixel/sample format.
Definition: avcodec.h:1451
AV_LOG_DEBUG
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
Definition: log.h:201
decode.h
get_bits.h
DisplayHeight
@ DisplayHeight
Definition: cfhd.h:90
process_bayer
static void process_bayer(AVFrame *frame, int bpc)
Definition: cfhd.c:152
planes
static const struct @345 planes[]
CODEC_LONG_NAME
#define CODEC_LONG_NAME(str)
Definition: codec_internal.h:264
FrameType
FrameType
G723.1 frame types.
Definition: g723_1.h:63
if
if(ret)
Definition: filter_design.txt:179
AV_CODEC_CAP_FRAME_THREADS
#define AV_CODEC_CAP_FRAME_THREADS
Codec supports frame-level multithreading.
Definition: codec.h:113
FrameIndex
@ FrameIndex
Definition: cfhd.h:50
InputFormat
@ InputFormat
Definition: cfhd.h:81
dequant_and_decompand
static int dequant_and_decompand(CFHDContext *s, int level, int quantisation, int codebook)
Definition: cfhd.c:110
NULL
#define NULL
Definition: coverity.c:32
AVERROR_PATCHWELCOME
#define AVERROR_PATCHWELCOME
Not yet implemented in FFmpeg, patches welcome.
Definition: error.h:64
run
uint8_t run
Definition: svq3.c:203
CFHDContext::planes
int planes
Definition: cfhd.h:149
CFHDDSPContext::vert_filter
void(* vert_filter)(int16_t *output, ptrdiff_t out_stride, const int16_t *low, ptrdiff_t low_stride, const int16_t *high, ptrdiff_t high_stride, int width, int height)
Definition: cfhddsp.h:31
AVCodecContext::internal
struct AVCodecInternal * internal
Private context used for internal data.
Definition: avcodec.h:433
CFHDDSPContext::horiz_filter
void(* horiz_filter)(int16_t *output, ptrdiff_t out_stride, const int16_t *low, ptrdiff_t low_stride, const int16_t *high, ptrdiff_t high_stride, int width, int height)
Definition: cfhddsp.h:26
Peak
Definition: cfhd.h:133
abs
#define abs(x)
Definition: cuda_runtime.h:35
UPDATE_THREAD_CONTEXT
#define UPDATE_THREAD_CONTEXT(func)
Definition: codec_internal.h:273
AV_PIX_FMT_YUV422P10
#define AV_PIX_FMT_YUV422P10
Definition: pixfmt.h:460
VLC_BITS
#define VLC_BITS
Definition: cfhd.h:95
HighpassHeight
@ HighpassHeight
Definition: cfhd.h:63
free_buffers
static void free_buffers(CFHDContext *s)
Definition: cfhd.c:220
bytestream2_get_bytes_left
static av_always_inline int bytestream2_get_bytes_left(GetByteContext *g)
Definition: bytestream.h:158
bytestream2_tell
static av_always_inline int bytestream2_tell(GetByteContext *g)
Definition: bytestream.h:192
SUBBAND_COUNT
#define SUBBAND_COUNT
Definition: cfhd.h:96
alloc_buffers
static int alloc_buffers(AVCodecContext *avctx)
Definition: cfhd.c:247
AV_CODEC_CAP_DR1
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
Definition: codec.h:52
Peak::base
GetByteContext base
Definition: cfhd.h:136
AVPacket::size
int size
Definition: packet.h:375
VersionRevision
@ VersionRevision
Definition: cfhd.h:37
codec_internal.h
for
for(k=2;k<=8;++k)
Definition: h264pred_template.c:425
BitstreamMarker
@ BitstreamMarker
Definition: cfhd.h:34
cfhd_decode
static int cfhd_decode(AVCodecContext *avctx, AVFrame *pic, int *got_frame, AVPacket *avpkt)
Definition: cfhd.c:376
avpriv_report_missing_feature
void avpriv_report_missing_feature(void *avc, const char *msg,...) av_printf_format(2
Log a generic warning message about a missing feature.
buffer.h
OPEN_READER
#define OPEN_READER(name, gb)
Definition: get_bits.h:138
height
#define height
CFHDContext::a_width
int a_width
Definition: cfhd.h:160
CFHDContext::coded_height
int coded_height
Definition: cfhd.h:155
offset
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
Definition: writing_filters.txt:86
attributes.h
VersionEdit
@ VersionEdit
Definition: cfhd.h:38
AVFrame::interlaced_frame
int interlaced_frame
The content of the picture is interlaced.
Definition: frame.h:477
CFHDContext::coded_format
enum AVPixelFormat coded_format
Definition: cfhd.h:157
CFHDContext::transform_type
int transform_type
Definition: cfhd.h:153
i
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:269
SubbandCount
@ SubbandCount
Definition: cfhd.h:43
CFHDContext::plane
Plane plane[4]
Definition: cfhd.h:180
AV_PIX_FMT_GBRP12
#define AV_PIX_FMT_GBRP12
Definition: pixfmt.h:476
av_malloc_array
#define av_malloc_array(a, b)
Definition: tableprint_vlc.h:31
CFHDContext
Definition: cfhd.h:139
common.h
SubbandBand
@ SubbandBand
Definition: cfhd.h:71
SubBand::read_ok
int8_t read_ok
Definition: cfhd.h:114
AVCodec::name
const char * name
Name of the codec implementation.
Definition: codec.h:211
update_thread_context
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before as well as code calling up to before the decode process starts Call have update_thread_context() run it in the next thread. Add AV_CODEC_CAP_FRAME_THREADS to the codec capabilities. There will be very little speed gain at this point but it should work. If there are inter-frame dependencies
SUBBAND_COUNT_3D
#define SUBBAND_COUNT_3D
Definition: cfhd.h:97
AVCodecContext::height
int height
Definition: avcodec.h:571
AVCodecContext::pix_fmt
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
Definition: avcodec.h:608
ImageWidth
@ ImageWidth
Definition: cfhd.h:48
av_calloc
void * av_calloc(size_t nmemb, size_t size)
Definition: mem.c:272
avcodec.h
stride
#define stride
Definition: h264pred_template.c:537
GET_RL_VLC
#define GET_RL_VLC(level, run, name, gb, table, bits, max_depth, need_update)
Definition: get_bits.h:728
tag
uint32_t tag
Definition: movenc.c:1647
ret
ret
Definition: filter_design.txt:187
CFHDContext::coded_width
int coded_width
Definition: cfhd.h:154
frame
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
Definition: filter_design.txt:264
inverse_temporal_filter
static void inverse_temporal_filter(int16_t *low, int16_t *high, int width)
Definition: cfhd.c:209
ff_thread_finish_setup
the pkt_dts and pkt_pts fields in AVFrame will work as usual Restrictions on codec whose streams don t reset across will not work because their bitstreams cannot be decoded in parallel *The contents of buffers must not be read before as well as code calling up to before the decode process starts Call ff_thread_finish_setup() afterwards. If some code can 't be moved
PeakLevel
@ PeakLevel
Definition: cfhd.h:83
BandCodingFlags
@ BandCodingFlags
Definition: cfhd.h:82
U
#define U(x)
Definition: vpx_arith.h:37
interlaced_vertical_filter
static void interlaced_vertical_filter(int16_t *output, int16_t *low, int16_t *high, int width, int linesize, int plane)
Definition: cfhd.c:196
EncodedFormat
@ EncodedFormat
Definition: cfhd.h:89
AVCodecContext
main external API structure.
Definition: avcodec.h:398
AVFrame::height
int height
Definition: frame.h:397
init_plane_defaults
static void init_plane_defaults(CFHDContext *s)
Definition: cfhd.c:74
AV_PIX_FMT_NONE
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:65
CFHDContext::a_transform_type
int a_transform_type
Definition: cfhd.h:163
ff_cfhddsp_init
av_cold void ff_cfhddsp_init(CFHDDSPContext *c, int depth, int bayer)
Definition: cfhddsp.c:105
ImageHeight
@ ImageHeight
Definition: cfhd.h:49
Plane
Definition: cfhd.h:117
BandWidth
@ BandWidth
Definition: cfhd.h:69
factor
static const int factor[16]
Definition: vf_pp7.c:76
SampleFlags
@ SampleFlags
Definition: cfhd.h:78
AVCodecContext::coded_width
int coded_width
Bitstream width / height, may be different from width/height e.g.
Definition: avcodec.h:586
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
Plane::idwt_buf
int16_t * idwt_buf
Definition: cfhd.h:122
DWT_LEVELS_3D
#define DWT_LEVELS_3D
Definition: cfhd.h:106
Plane::band
SubBand band[DWT_LEVELS_3D][4]
Definition: cfhd.h:130
LowpassHeight
@ LowpassHeight
Definition: cfhd.h:54
FFALIGN
#define FFALIGN(x, a)
Definition: macros.h:78
alpha
static const int16_t alpha[]
Definition: ilbcdata.h:55
AVPacket
This structure stores compressed data.
Definition: packet.h:351
AVCodecContext::priv_data
void * priv_data
Definition: avcodec.h:425
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:34
src
INIT_CLIP pixel * src
Definition: h264pred_template.c:418
bytestream.h
bytestream2_init
static av_always_inline void bytestream2_init(GetByteContext *g, const uint8_t *buf, int buf_size)
Definition: bytestream.h:137
AVFrame::linesize
int linesize[AV_NUM_DATA_POINTERS]
For video, a positive or negative value, which is typically indicating the size in bytes of each pict...
Definition: frame.h:370
coeff
static const double coeff[2][5]
Definition: vf_owdenoise.c:78
av_log
#define av_log(a,...)
Definition: tableprint_vlc.h:27
AVERROR_INVALIDDATA
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
Definition: error.h:61
AV_PIX_FMT_BAYER_RGGB16
#define AV_PIX_FMT_BAYER_RGGB16
Definition: pixfmt.h:484
SampleIndexTable
@ SampleIndexTable
Definition: cfhd.h:33
cfhd_close
static av_cold int cfhd_close(AVCodecContext *avctx)
Definition: cfhd.c:1401
init_frame_defaults
static void init_frame_defaults(CFHDContext *s)
Definition: cfhd.c:88
channel
channel
Definition: ebur128.h:39
PeakOffsetLow
@ PeakOffsetLow
Definition: cfhd.h:84
codebook
static const unsigned codebook[256][2]
Definition: cfhdenc.c:42
CFHDContext::a_height
int a_height
Definition: cfhd.h:161
re
float re
Definition: fft.c:79