FFmpeg
vf_deband.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2015 Niklas Haas
3  * Copyright (c) 2015 Paul B Mahol
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a copy
6  * of this software and associated documentation files (the "Software"), to deal
7  * in the Software without restriction, including without limitation the rights
8  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9  * copies of the Software, and to permit persons to whom the Software is
10  * furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice shall be included in
13  * all copies or substantial portions of the Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21  * SOFTWARE.
22  */
23 
24 #include "libavutil/opt.h"
25 #include "libavutil/pixdesc.h"
26 #include "avfilter.h"
27 #include "internal.h"
28 #include "video.h"
29 
30 typedef struct DebandContext {
31  const AVClass *class;
32 
33  int coupling;
34  float threshold[4];
35  int range;
36  int blur;
37  float direction;
38 
40  int planewidth[4];
41  int planeheight[4];
42  int shift[2];
43  int thr[4];
44 
45  int *x_pos;
46  int *y_pos;
47 
48  int (*deband)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
50 
51 #define OFFSET(x) offsetof(DebandContext, x)
52 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
53 
54 static const AVOption deband_options[] = {
55  { "1thr", "set 1st plane threshold", OFFSET(threshold[0]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0.00003, 0.5, FLAGS },
56  { "2thr", "set 2nd plane threshold", OFFSET(threshold[1]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0.00003, 0.5, FLAGS },
57  { "3thr", "set 3rd plane threshold", OFFSET(threshold[2]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0.00003, 0.5, FLAGS },
58  { "4thr", "set 4th plane threshold", OFFSET(threshold[3]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0.00003, 0.5, FLAGS },
59  { "range", "set range", OFFSET(range), AV_OPT_TYPE_INT, {.i64=16}, INT_MIN, INT_MAX, FLAGS },
60  { "r", "set range", OFFSET(range), AV_OPT_TYPE_INT, {.i64=16}, INT_MIN, INT_MAX, FLAGS },
61  { "direction", "set direction", OFFSET(direction), AV_OPT_TYPE_FLOAT, {.dbl=2*M_PI},-2*M_PI, 2*M_PI, FLAGS },
62  { "d", "set direction", OFFSET(direction), AV_OPT_TYPE_FLOAT, {.dbl=2*M_PI},-2*M_PI, 2*M_PI, FLAGS },
63  { "blur", "set blur", OFFSET(blur), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
64  { "b", "set blur", OFFSET(blur), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
65  { "coupling", "set plane coupling", OFFSET(coupling), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS },
66  { "c", "set plane coupling", OFFSET(coupling), AV_OPT_TYPE_BOOL, {.i64=0}, 0, 1, FLAGS },
67  { NULL }
68 };
69 
71 
73 {
74  DebandContext *s = ctx->priv;
75 
76  static const enum AVPixelFormat pix_fmts[] = {
96  };
97 
98  static const enum AVPixelFormat cpix_fmts[] = {
108  AV_PIX_FMT_NONE
109  };
110 
111  AVFilterFormats *fmts_list = ff_make_format_list(s->coupling ? cpix_fmts : pix_fmts);
112  if (!fmts_list)
113  return AVERROR(ENOMEM);
114 
115  return ff_set_common_formats(ctx, fmts_list);
116 }
117 
118 static float frand(int x, int y)
119 {
120  const float r = sinf(x * 12.9898 + y * 78.233) * 43758.545;
121 
122  return r - floorf(r);
123 }
124 
125 static int inline get_avg(int ref0, int ref1, int ref2, int ref3)
126 {
127  return (ref0 + ref1 + ref2 + ref3) / 4;
128 }
129 
130 typedef struct ThreadData {
131  AVFrame *in, *out;
132 } ThreadData;
133 
134 static int deband_8_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
135 {
136  DebandContext *s = ctx->priv;
137  ThreadData *td = arg;
138  AVFrame *in = td->in;
139  AVFrame *out = td->out;
140  int x, y, p;
141 
142  for (p = 0; p < s->nb_components; p++) {
143  const uint8_t *src_ptr = (const uint8_t *)in->data[p];
144  uint8_t *dst_ptr = (uint8_t *)out->data[p];
145  const int dst_linesize = out->linesize[p];
146  const int src_linesize = in->linesize[p];
147  const int thr = s->thr[p];
148  const int start = (s->planeheight[p] * jobnr ) / nb_jobs;
149  const int end = (s->planeheight[p] * (jobnr+1)) / nb_jobs;
150  const int w = s->planewidth[p] - 1;
151  const int h = s->planeheight[p] - 1;
152 
153  for (y = start; y < end; y++) {
154  const int pos = y * s->planewidth[0];
155 
156  for (x = 0; x < s->planewidth[p]; x++) {
157  const int x_pos = s->x_pos[pos + x];
158  const int y_pos = s->y_pos[pos + x];
159  const int ref0 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
160  const int ref1 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
161  const int ref2 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
162  const int ref3 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
163  const int src0 = src_ptr[y * src_linesize + x];
164 
165  if (s->blur) {
166  const int avg = get_avg(ref0, ref1, ref2, ref3);
167  const int diff = FFABS(src0 - avg);
168 
169  dst_ptr[y * dst_linesize + x] = diff < thr ? avg : src0;
170  } else {
171  dst_ptr[y * dst_linesize + x] = (FFABS(src0 - ref0) < thr) &&
172  (FFABS(src0 - ref1) < thr) &&
173  (FFABS(src0 - ref2) < thr) &&
174  (FFABS(src0 - ref3) < thr) ? get_avg(ref0, ref1, ref2, ref3) : src0;
175  }
176  }
177  }
178  }
179 
180  return 0;
181 }
182 
183 static int deband_8_coupling_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
184 {
185  DebandContext *s = ctx->priv;
186  ThreadData *td = arg;
187  AVFrame *in = td->in;
188  AVFrame *out = td->out;
189  const int start = (s->planeheight[0] * jobnr ) / nb_jobs;
190  const int end = (s->planeheight[0] * (jobnr+1)) / nb_jobs;
191  int x, y, p;
192 
193  for (y = start; y < end; y++) {
194  const int pos = y * s->planewidth[0];
195 
196  for (x = 0; x < s->planewidth[0]; x++) {
197  const int x_pos = s->x_pos[pos + x];
198  const int y_pos = s->y_pos[pos + x];
199  int avg[4], cmp[4] = { 0 }, src[4];
200 
201  for (p = 0; p < s->nb_components; p++) {
202  const uint8_t *src_ptr = (const uint8_t *)in->data[p];
203  const int src_linesize = in->linesize[p];
204  const int thr = s->thr[p];
205  const int w = s->planewidth[p] - 1;
206  const int h = s->planeheight[p] - 1;
207  const int ref0 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
208  const int ref1 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
209  const int ref2 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
210  const int ref3 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
211  const int src0 = src_ptr[y * src_linesize + x];
212 
213  src[p] = src0;
214  avg[p] = get_avg(ref0, ref1, ref2, ref3);
215 
216  if (s->blur) {
217  cmp[p] = FFABS(src0 - avg[p]) < thr;
218  } else {
219  cmp[p] = (FFABS(src0 - ref0) < thr) &&
220  (FFABS(src0 - ref1) < thr) &&
221  (FFABS(src0 - ref2) < thr) &&
222  (FFABS(src0 - ref3) < thr);
223  }
224  }
225 
226  for (p = 0; p < s->nb_components; p++)
227  if (!cmp[p])
228  break;
229  if (p == s->nb_components) {
230  for (p = 0; p < s->nb_components; p++) {
231  const int dst_linesize = out->linesize[p];
232 
233  out->data[p][y * dst_linesize + x] = avg[p];
234  }
235  } else {
236  for (p = 0; p < s->nb_components; p++) {
237  const int dst_linesize = out->linesize[p];
238 
239  out->data[p][y * dst_linesize + x] = src[p];
240  }
241  }
242  }
243  }
244 
245  return 0;
246 }
247 
248 static int deband_16_coupling_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
249 {
250  DebandContext *s = ctx->priv;
251  ThreadData *td = arg;
252  AVFrame *in = td->in;
253  AVFrame *out = td->out;
254  const int start = (s->planeheight[0] * jobnr ) / nb_jobs;
255  const int end = (s->planeheight[0] * (jobnr+1)) / nb_jobs;
256  int x, y, p, z;
257 
258  for (y = start; y < end; y++) {
259  const int pos = y * s->planewidth[0];
260 
261  for (x = 0; x < s->planewidth[0]; x++) {
262  const int x_pos = s->x_pos[pos + x];
263  const int y_pos = s->y_pos[pos + x];
264  int avg[4], cmp[4] = { 0 }, src[4];
265 
266  for (p = 0; p < s->nb_components; p++) {
267  const uint16_t *src_ptr = (const uint16_t *)in->data[p];
268  const int src_linesize = in->linesize[p] / 2;
269  const int thr = s->thr[p];
270  const int w = s->planewidth[p] - 1;
271  const int h = s->planeheight[p] - 1;
272  const int ref0 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
273  const int ref1 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
274  const int ref2 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
275  const int ref3 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
276  const int src0 = src_ptr[y * src_linesize + x];
277 
278  src[p] = src0;
279  avg[p] = get_avg(ref0, ref1, ref2, ref3);
280 
281  if (s->blur) {
282  cmp[p] = FFABS(src0 - avg[p]) < thr;
283  } else {
284  cmp[p] = (FFABS(src0 - ref0) < thr) &&
285  (FFABS(src0 - ref1) < thr) &&
286  (FFABS(src0 - ref2) < thr) &&
287  (FFABS(src0 - ref3) < thr);
288  }
289  }
290 
291  for (z = 0; z < s->nb_components; z++)
292  if (!cmp[z])
293  break;
294  if (z == s->nb_components) {
295  for (p = 0; p < s->nb_components; p++) {
296  const int dst_linesize = out->linesize[p] / 2;
297  uint16_t *dst = (uint16_t *)out->data[p] + y * dst_linesize + x;
298 
299  dst[0] = avg[p];
300  }
301  } else {
302  for (p = 0; p < s->nb_components; p++) {
303  const int dst_linesize = out->linesize[p] / 2;
304  uint16_t *dst = (uint16_t *)out->data[p] + y * dst_linesize + x;
305 
306  dst[0] = src[p];
307  }
308  }
309  }
310  }
311 
312  return 0;
313 }
314 
315 static int deband_16_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
316 {
317  DebandContext *s = ctx->priv;
318  ThreadData *td = arg;
319  AVFrame *in = td->in;
320  AVFrame *out = td->out;
321  int x, y, p;
322 
323  for (p = 0; p < s->nb_components; p++) {
324  const uint16_t *src_ptr = (const uint16_t *)in->data[p];
325  uint16_t *dst_ptr = (uint16_t *)out->data[p];
326  const int dst_linesize = out->linesize[p] / 2;
327  const int src_linesize = in->linesize[p] / 2;
328  const int thr = s->thr[p];
329  const int start = (s->planeheight[p] * jobnr ) / nb_jobs;
330  const int end = (s->planeheight[p] * (jobnr+1)) / nb_jobs;
331  const int w = s->planewidth[p] - 1;
332  const int h = s->planeheight[p] - 1;
333 
334  for (y = start; y < end; y++) {
335  const int pos = y * s->planewidth[0];
336 
337  for (x = 0; x < s->planewidth[p]; x++) {
338  const int x_pos = s->x_pos[pos + x];
339  const int y_pos = s->y_pos[pos + x];
340  const int ref0 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
341  const int ref1 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + x_pos, 0, w)];
342  const int ref2 = src_ptr[av_clip(y + -y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
343  const int ref3 = src_ptr[av_clip(y + y_pos, 0, h) * src_linesize + av_clip(x + -x_pos, 0, w)];
344  const int src0 = src_ptr[y * src_linesize + x];
345 
346  if (s->blur) {
347  const int avg = get_avg(ref0, ref1, ref2, ref3);
348  const int diff = FFABS(src0 - avg);
349 
350  dst_ptr[y * dst_linesize + x] = diff < thr ? avg : src0;
351  } else {
352  dst_ptr[y * dst_linesize + x] = (FFABS(src0 - ref0) < thr) &&
353  (FFABS(src0 - ref1) < thr) &&
354  (FFABS(src0 - ref2) < thr) &&
355  (FFABS(src0 - ref3) < thr) ? get_avg(ref0, ref1, ref2, ref3) : src0;
356  }
357  }
358  }
359  }
360 
361  return 0;
362 }
363 
365 {
367  AVFilterContext *ctx = inlink->dst;
368  DebandContext *s = ctx->priv;
369  const float direction = s->direction;
370  const int range = s->range;
371  int x, y;
372 
373  s->nb_components = desc->nb_components;
374 
375  s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
376  s->planeheight[0] = s->planeheight[3] = inlink->h;
377  s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w);
378  s->planewidth[0] = s->planewidth[3] = inlink->w;
379  s->shift[0] = desc->log2_chroma_w;
380  s->shift[1] = desc->log2_chroma_h;
381 
382  if (s->coupling)
384  else
385  s->deband = desc->comp[0].depth > 8 ? deband_16_c : deband_8_c;
386 
387  s->thr[0] = ((1 << desc->comp[0].depth) - 1) * s->threshold[0];
388  s->thr[1] = ((1 << desc->comp[1].depth) - 1) * s->threshold[1];
389  s->thr[2] = ((1 << desc->comp[2].depth) - 1) * s->threshold[2];
390  s->thr[3] = ((1 << desc->comp[3].depth) - 1) * s->threshold[3];
391 
392  s->x_pos = av_malloc(s->planewidth[0] * s->planeheight[0] * sizeof(*s->x_pos));
393  s->y_pos = av_malloc(s->planewidth[0] * s->planeheight[0] * sizeof(*s->y_pos));
394  if (!s->x_pos || !s->y_pos)
395  return AVERROR(ENOMEM);
396 
397  for (y = 0; y < s->planeheight[0]; y++) {
398  for (x = 0; x < s->planewidth[0]; x++) {
399  const float r = frand(x, y);
400  const float dir = direction < 0 ? -direction : r * direction;
401  const int dist = range < 0 ? -range : r * range;
402 
403  s->x_pos[y * s->planewidth[0] + x] = cosf(dir) * dist;
404  s->y_pos[y * s->planewidth[0] + x] = sinf(dir) * dist;
405  }
406  }
407 
408  return 0;
409 }
410 
412 {
413  AVFilterContext *ctx = inlink->dst;
414  AVFilterLink *outlink = ctx->outputs[0];
415  DebandContext *s = ctx->priv;
416  AVFrame *out;
417  ThreadData td;
418 
419  out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
420  if (!out) {
421  av_frame_free(&in);
422  return AVERROR(ENOMEM);
423  }
424  av_frame_copy_props(out, in);
425 
426  td.in = in; td.out = out;
427  ctx->internal->execute(ctx, s->deband, &td, NULL, FFMIN3(s->planeheight[1],
428  s->planeheight[2],
430 
431  av_frame_free(&in);
432  return ff_filter_frame(outlink, out);
433 }
434 
436 {
437  DebandContext *s = ctx->priv;
438 
439  av_freep(&s->x_pos);
440  av_freep(&s->y_pos);
441 }
442 
444  {
445  .name = "default",
446  .type = AVMEDIA_TYPE_VIDEO,
447  .config_props = config_input,
448  .filter_frame = filter_frame,
449  },
450  { NULL }
451 };
452 
454  {
455  .name = "default",
456  .type = AVMEDIA_TYPE_VIDEO,
457  },
458  { NULL }
459 };
460 
462  .name = "deband",
463  .description = NULL_IF_CONFIG_SMALL("Debands video."),
464  .priv_size = sizeof(DebandContext),
465  .priv_class = &deband_class,
466  .uninit = uninit,
468  .inputs = avfilter_vf_deband_inputs,
469  .outputs = avfilter_vf_deband_outputs,
471 };
#define NULL
Definition: coverity.c:32
#define AV_PIX_FMT_YUVA422P16
Definition: pixfmt.h:430
static const AVFilterPad avfilter_vf_deband_outputs[]
Definition: vf_deband.c:453
AVFrame * out
Definition: af_adeclick.c:488
#define AV_PIX_FMT_YUVA422P9
Definition: pixfmt.h:422
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2522
This structure describes decoded (raw) audio or video data.
Definition: frame.h:295
AVOption.
Definition: opt.h:246
#define AV_PIX_FMT_YUVA420P10
Definition: pixfmt.h:424
#define AV_PIX_FMT_YUV444P14
Definition: pixfmt.h:397
#define AV_PIX_FMT_YUVA422P10
Definition: pixfmt.h:425
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:71
Main libavfilter public API header.
const char * desc
Definition: nvenc.c:68
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:168
#define AV_PIX_FMT_GBRP10
Definition: pixfmt.h:403
#define AV_PIX_FMT_GRAY9
Definition: pixfmt.h:367
#define AV_PIX_FMT_YUV420P12
Definition: pixfmt.h:391
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
Definition: video.c:99
#define src
Definition: vp8dsp.c:254
static int query_formats(AVFilterContext *ctx)
Definition: vf_deband.c:72
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
Definition: pixdesc.h:92
AVFilterFormats * ff_make_format_list(const int *fmts)
Create a list of supported formats.
Definition: formats.c:283
#define AV_PIX_FMT_GRAY10
Definition: pixfmt.h:368
AVFILTER_DEFINE_CLASS(deband)
#define AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
Some filters support a generic "enable" expression option that can be used to enable or disable a fil...
Definition: avfilter.h:125
const char * name
Pad name.
Definition: internal.h:60
#define AV_PIX_FMT_GRAY12
Definition: pixfmt.h:369
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1093
int thr[4]
Definition: vf_deband.c:43
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:101
static const AVOption deband_options[]
Definition: vf_deband.c:54
AVComponentDescriptor comp[4]
Parameters that describe how pixels are packed.
Definition: pixdesc.h:117
uint8_t
#define av_cold
Definition: attributes.h:82
#define av_malloc(s)
AVOptions.
static av_cold int end(AVCodecContext *avctx)
Definition: avrndec.c:90
static int deband_8_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_deband.c:134
#define AV_PIX_FMT_YUVA420P9
Definition: pixfmt.h:421
#define cosf(x)
Definition: libm.h:78
#define AV_PIX_FMT_GBRP9
Definition: pixfmt.h:402
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range...
Definition: pixfmt.h:100
#define FFMIN3(a, b, c)
Definition: common.h:97
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:79
#define AV_PIX_FMT_YUV444P16
Definition: pixfmt.h:400
#define AV_PIX_FMT_YUV422P12
Definition: pixfmt.h:392
#define AV_PIX_FMT_YUVA420P16
Definition: pixfmt.h:429
A filter pad used for either input or output.
Definition: internal.h:54
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:176
float threshold[4]
Definition: vf_deband.c:34
static av_cold void uninit(AVFilterContext *ctx)
Definition: vf_deband.c:435
int ff_set_common_formats(AVFilterContext *ctx, AVFilterFormats *formats)
A helper for query_formats() which sets all links to the same list of formats.
Definition: formats.c:569
#define td
Definition: regdef.h:70
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
Definition: pixdesc.h:101
#define OFFSET(x)
Definition: vf_deband.c:51
int(* deband)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_deband.c:48
static float frand(int x, int y)
Definition: vf_deband.c:118
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:202
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
Definition: internal.h:186
const char * r
Definition: vf_curves.c:114
void * priv
private data for use by the filter
Definition: avfilter.h:353
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
Definition: avfilter.h:116
#define AV_PIX_FMT_YUVA444P16
Definition: pixfmt.h:431
const char * arg
Definition: jacosubdec.c:66
static int deband_8_coupling_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_deband.c:183
#define AV_PIX_FMT_GBRAP16
Definition: pixfmt.h:409
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:70
#define AV_PIX_FMT_YUV422P9
Definition: pixfmt.h:385
uint8_t nb_components
The number of components each pixel has, (1-4)
Definition: pixdesc.h:83
#define AV_PIX_FMT_GBRP16
Definition: pixfmt.h:406
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
Definition: avfilter.c:802
#define AV_PIX_FMT_GRAY16
Definition: pixfmt.h:371
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
static int deband_16_coupling_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_deband.c:248
uint8_t w
Definition: llviddspenc.c:38
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
Definition: vf_deband.c:411
AVFormatContext * ctx
Definition: movenc.c:48
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
Definition: common.h:72
#define s(width, name)
Definition: cbs_vp9.c:257
int nb_components
Definition: vf_deband.c:39
static av_always_inline int cmp(MpegEncContext *s, const int x, const int y, const int subx, const int suby, const int size, const int h, int ref_index, int src_index, me_cmp_func cmp_func, me_cmp_func chroma_cmp_func, const int flags)
compares a block (either a full macroblock or a partition thereof) against a proposed motion-compensa...
Definition: motion_est.c:260
#define AV_PIX_FMT_YUVA444P10
Definition: pixfmt.h:426
#define AV_PIX_FMT_YUV444P9
Definition: pixfmt.h:386
#define AV_PIX_FMT_GBRP14
Definition: pixfmt.h:405
static const AVFilterPad outputs[]
Definition: af_acontrast.c:203
#define AV_PIX_FMT_YUV420P16
Definition: pixfmt.h:398
#define sinf(x)
Definition: libm.h:419
#define AV_PIX_FMT_YUV420P14
Definition: pixfmt.h:395
int planewidth[4]
Definition: vf_deband.c:40
Used for passing data between threads.
Definition: dsddec.c:64
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:326
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:177
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:81
#define AV_PIX_FMT_GRAY14
Definition: pixfmt.h:370
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several inputs
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi-0x80)*(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi-0x80)*(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t,*(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t,*(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t,*(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t,*(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31))))#define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac){}void ff_audio_convert_free(AudioConvert **ac){if(!*ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);}AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map){AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method!=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2){ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc){av_free(ac);return NULL;}return ac;}in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar){ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar?ac->channels:1;}else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;}int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in){int use_generic=1;int len=in->nb_samples;int p;if(ac->dc){av_log(ac->avr, AV_LOG_TRACE,"%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
Describe the class of an AVClass context structure.
Definition: log.h:67
Filter definition.
Definition: avfilter.h:144
int * y_pos
Definition: vf_deband.c:46
#define src0
Definition: h264pred.c:138
const char * name
Filter name.
Definition: avfilter.h:148
#define FLAGS
Definition: vf_deband.c:52
#define AV_PIX_FMT_YUV420P9
Definition: pixfmt.h:384
AVFilterLink ** outputs
array of pointers to output links
Definition: avfilter.h:350
static enum AVPixelFormat pix_fmts[]
Definition: libkvazaar.c:275
int * x_pos
Definition: vf_deband.c:45
#define AV_PIX_FMT_YUV422P14
Definition: pixfmt.h:396
#define AV_PIX_FMT_GBRP12
Definition: pixfmt.h:404
#define flags(name, subs,...)
Definition: cbs_av1.c:561
AVFilterInternal * internal
An opaque struct for libavfilter internal use.
Definition: avfilter.h:378
#define AV_PIX_FMT_YUV444P12
Definition: pixfmt.h:394
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:309
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
#define avg(a, b, c, d)
int
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:66
float direction
Definition: vf_deband.c:37
Y , 8bpp.
Definition: pixfmt.h:74
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:215
#define AV_PIX_FMT_YUVA444P9
Definition: pixfmt.h:423
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:80
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:73
avfilter_execute_func * execute
Definition: internal.h:155
AVFilter ff_vf_deband
Definition: vf_deband.c:461
int shift[2]
Definition: vf_deband.c:42
static av_always_inline int diff(const uint32_t a, const uint32_t b)
static int config_input(AVFilterLink *inlink)
Definition: vf_deband.c:364
int planeheight[4]
Definition: vf_deband.c:41
A list of supported formats for one end of a filter link.
Definition: formats.h:64
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:258
static const AVFilterPad avfilter_vf_deband_inputs[]
Definition: vf_deband.c:443
An instance of a filter.
Definition: avfilter.h:338
FILE * out
Definition: movenc.c:54
static int get_avg(int ref0, int ref1, int ref2, int ref3)
Definition: vf_deband.c:125
#define av_freep(p)
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:99
void INT64 start
Definition: avisynth_c.h:766
#define M_PI
Definition: mathematics.h:52
AVFrame * in
Definition: af_afftdn.c:1082
internal API functions
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
int depth
Number of bits in the component.
Definition: pixdesc.h:58
static int deband_16_c(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_deband.c:315
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
#define AV_PIX_FMT_YUV422P16
Definition: pixfmt.h:399
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Definition: frame.c:654
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:58