FFmpeg
vf_yaepblur.c
Go to the documentation of this file.
1 /*
2  * Copyright (C) 2019 Leo Zhang <leozhang@qiyi.com>
3 
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 /**
22  * @file
23  * yaep(yet another edge preserving) blur filter
24  *
25  * This implementation is based on an algorithm described in
26  * "J. S. Lee, Digital image enhancement and noise filtering by use of local statistics, IEEE Trans. Pattern
27  * Anal. Mach. Intell. PAMI-2, 1980."
28  */
29 
30 #include "libavutil/opt.h"
31 #include "libavutil/imgutils.h"
32 #include "avfilter.h"
33 #include "internal.h"
34 
35 typedef struct YAEPContext {
36  const AVClass *class;
37 
38  int planes;
39  int radius;
40  int sigma;
41 
42  int nb_planes;
43  int planewidth[4];
44  int planeheight[4];
45  int depth;
46 
47  uint64_t *sat; ///< summed area table
48  uint64_t *square_sat; ///< square summed area table
50 
51  int (*pre_calculate_row)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
52  int (*filter_slice )(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
53 } YAEPContext;
54 
56 {
57  YAEPContext *s = ctx->priv;
58  av_freep(&s->sat);
59  av_freep(&s->square_sat);
60 }
61 
63 {
64  static const enum AVPixelFormat pix_fmts[] = {
84  };
85 
86  return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
87 }
88 
89 typedef struct ThreadData {
90  int width;
91  int height;
92  int src_linesize;
93  int dst_linesize;
94  uint8_t *src;
95  uint8_t *dst;
96 } ThreadData;
97 
98 #define PRE_CALCULATE_ROW(type, name) \
99 static int pre_calculate_row_##name(AVFilterContext *ctx, void *arg, \
100  int jobnr, int nb_jobs) \
101 { \
102  ThreadData *td = arg; \
103  YAEPContext *s = ctx->priv; \
104  \
105  const int width = td->width; \
106  const int height = td->height; \
107  const int linesize = td->src_linesize / sizeof(type); \
108  const int sat_linesize = s->sat_linesize; \
109  \
110  const int starty = height * jobnr / nb_jobs; \
111  const int endy = height * (jobnr+1) / nb_jobs; \
112  \
113  uint64_t *sat = s->sat + (starty + 1) * sat_linesize; \
114  uint64_t *square_sat = s->square_sat + (starty + 1) * sat_linesize; \
115  const type *src = (const type *)td->src + starty * linesize; \
116  \
117  int x, y; \
118  \
119  for (y = starty; y < endy; y++) { \
120  for (x = 0; x < width; x++) { \
121  sat[x+1] = sat[x] + src[x]; \
122  square_sat[x+1] = square_sat[x] + (uint64_t)src[x] * src[x]; \
123  } \
124  sat += sat_linesize; \
125  square_sat += sat_linesize; \
126  src += linesize; \
127  } \
128  \
129  return 0; \
130 }
131 
133 PRE_CALCULATE_ROW(uint16_t, word)
134 
136  int jobnr, int nb_jobs)
137 {
138  ThreadData *td = arg;
139  YAEPContext *s = ctx->priv;
140 
141  const int width = td->width;
142  const int height = td->height;
143  const int sat_linesize = s->sat_linesize;
144 
145  const int startx = width * jobnr / nb_jobs;
146  const int endx = width * (jobnr + 1) / nb_jobs;
147 
148  uint64_t *sat, *square_sat;
149  int x, y;
150 
151  for (x = startx; x < endx; x++) {
152  sat = s->sat + x + 1;
153  square_sat = s->square_sat + x + 1;
154  for (y = 0; y < height; y++) {
155  *(sat+sat_linesize) += *sat;
156  *(square_sat+sat_linesize) += *square_sat;
157  sat += sat_linesize;
158  square_sat += sat_linesize;
159  }
160  }
161 
162  return 0;
163 }
164 
165 #define FILTER_SLICE(type, name) \
166 static int filter_slice_##name(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) \
167 { \
168  ThreadData *td = arg; \
169  YAEPContext *s = ctx->priv; \
170  \
171  const int width = td->width; \
172  const int height = td->height; \
173  const int src_linesize = td->src_linesize / sizeof(type); \
174  const int dst_linesize = td->dst_linesize / sizeof(type); \
175  const int sat_linesize = s->sat_linesize; \
176  const int sigma = s->sigma; \
177  const int radius = s->radius; \
178  \
179  uint64_t *sat = s->sat; \
180  uint64_t *square_sat = s->square_sat; \
181  const type *src = (const type *)td->src; \
182  type *dst = (type *)td->dst; \
183  \
184  const int starty = height * jobnr / nb_jobs; \
185  const int endy = height * (jobnr + 1) / nb_jobs; \
186  \
187  int x, y; \
188  int lower_x, higher_x; \
189  int lower_y, higher_y; \
190  int dist_y, count; \
191  uint64_t sum, square_sum, mean, var; \
192  \
193  for (y = starty; y < endy; y++) { \
194  lower_y = y - radius < 0 ? 0 : y - radius; \
195  higher_y = y + radius + 1 > height ? height : y + radius + 1; \
196  dist_y = higher_y - lower_y; \
197  for (x = 0; x < width; x++) { \
198  lower_x = x - radius < 0 ? 0 : x - radius; \
199  higher_x = x + radius + 1 > width ? width : x + radius + 1; \
200  count = dist_y * (higher_x - lower_x); \
201  sum = sat[higher_y * sat_linesize + higher_x] \
202  - sat[higher_y * sat_linesize + lower_x] \
203  - sat[lower_y * sat_linesize + higher_x] \
204  + sat[lower_y * sat_linesize + lower_x]; \
205  square_sum = square_sat[higher_y * sat_linesize + higher_x] \
206  - square_sat[higher_y * sat_linesize + lower_x] \
207  - square_sat[lower_y * sat_linesize + higher_x] \
208  + square_sat[lower_y * sat_linesize + lower_x]; \
209  mean = sum / count; \
210  var = (square_sum - sum * sum / count) / count; \
211  dst[y * dst_linesize + x] = (sigma * mean + var * src[y * src_linesize + x]) / (sigma + var); \
212  } \
213  } \
214  return 0; \
215 }
216 
218 FILTER_SLICE(uint16_t, word)
219 
221 {
222  AVFilterContext *ctx = inlink->dst;
223  YAEPContext *s = ctx->priv;
224  AVFilterLink *outlink = ctx->outputs[0];
225  AVFrame *out;
226  int plane;
227  const int nb_threads = ff_filter_get_nb_threads(ctx);
228  ThreadData td;
229 
230  if (av_frame_is_writable(in)) {
231  out = in;
232  } else {
233  out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
234  if (!out) {
235  av_frame_free(&in);
236  return AVERROR(ENOMEM);
237  }
238  av_frame_copy_props(out, in);
239  }
240 
241  for (plane = 0; plane < s->nb_planes; plane++) {
242  if (!s->radius || !(s->planes & (1<<plane))) {
243  if (out != in) {
244  av_image_copy_plane(out->data[plane], out->linesize[plane],
245  in->data[plane], in->linesize[plane],
246  s->planewidth[plane] * ((s->depth + 7) / 8),
247  s->planeheight[plane]);
248  }
249  continue;
250  }
251 
252  td.width = s->planewidth[plane];
253  td.height = s->planeheight[plane];
254  td.src = in->data[plane];
255  td.src_linesize = in->linesize[plane];
256  ctx->internal->execute(ctx, s->pre_calculate_row, &td, NULL, FFMIN(td.height, nb_threads));
257  ctx->internal->execute(ctx, pre_calculate_col, &td, NULL, FFMIN(td.width, nb_threads));
258 
259  td.dst = out->data[plane];
260  td.dst_linesize = out->linesize[plane];
261  ctx->internal->execute(ctx, s->filter_slice, &td, NULL, FFMIN(td.height, nb_threads));
262  }
263 
264  if (out != in)
265  av_frame_free(&in);
266 
267  return ff_filter_frame(outlink, out);
268 }
269 
271 {
272  YAEPContext *s = inlink->dst->priv;
274 
275  s->depth = desc->comp[0].depth;
276  s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w);
277  s->planewidth[0] = s->planewidth[3] = inlink->w;
278  s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
279  s->planeheight[0] = s->planeheight[3] = inlink->h;
281 
282  s->radius = FFMIN(s->radius, AV_CEIL_RSHIFT(FFMIN(inlink->w, inlink->h), 1));
283 
284  if (s->depth <= 8) {
285  s->pre_calculate_row = pre_calculate_row_byte;
286  s->filter_slice = filter_slice_byte;
287  } else {
288  s->pre_calculate_row = pre_calculate_row_word;
289  s->filter_slice = filter_slice_word;
290  }
291 
292  // padding one row on the top, and padding one col on the left, that is why + 1 below
293  s->sat_linesize = inlink->w + 1;
294  s->sat = av_mallocz_array(inlink->h + 1, s->sat_linesize*sizeof(*s->sat));
295  if (!s->sat)
296  return AVERROR(ENOMEM);
297 
298  s->square_sat = av_mallocz_array(inlink->h + 1, s->sat_linesize*sizeof(*s->square_sat));
299  if (!s->square_sat)
300  return AVERROR(ENOMEM);
301 
302  return 0;
303 }
304 
305 static const AVFilterPad yaep_inputs[] = {
306  {
307  .name = "default",
308  .type = AVMEDIA_TYPE_VIDEO,
309  .config_props = config_input,
310  .filter_frame = filter_frame,
311  },
312  { NULL }
313 };
314 
315 static const AVFilterPad yaep_outputs[] = {
316  {
317  .name = "default",
318  .type = AVMEDIA_TYPE_VIDEO,
319  },
320  { NULL }
321 };
322 
323 #define OFFSET(x) offsetof(YAEPContext, x)
324 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_RUNTIME_PARAM
325 
326 static const AVOption yaepblur_options[] = {
327  { "radius", "set window radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3}, 0, INT_MAX, .flags=FLAGS },
328  { "r" , "set window radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3}, 0, INT_MAX, .flags=FLAGS },
329  { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1}, 0, 0xF, .flags=FLAGS },
330  { "p", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1}, 0, 0xF, .flags=FLAGS },
331  { "sigma", "set blur strength", OFFSET(sigma), AV_OPT_TYPE_INT, {.i64=128}, 1, INT_MAX, .flags=FLAGS },
332  { "s", "set blur strength", OFFSET(sigma), AV_OPT_TYPE_INT, {.i64=128}, 1, INT_MAX, .flags=FLAGS },
333  { NULL }
334 };
335 
336 AVFILTER_DEFINE_CLASS(yaepblur);
337 
339  .name = "yaepblur",
340  .description = NULL_IF_CONFIG_SMALL("Yet another edge preserving blur filter."),
341  .priv_size = sizeof(YAEPContext),
342  .priv_class = &yaepblur_class,
343  .uninit = uninit,
345  .inputs = yaep_inputs,
346  .outputs = yaep_outputs,
349 };
#define NULL
Definition: coverity.c:32
#define AV_PIX_FMT_YUVA422P16
Definition: pixfmt.h:442
#define AV_PIX_FMT_YUVA422P9
Definition: pixfmt.h:434
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2573
This structure describes decoded (raw) audio or video data.
Definition: frame.h:308
AVOption.
Definition: opt.h:248
#define AV_PIX_FMT_YUVA420P10
Definition: pixfmt.h:436
#define AV_PIX_FMT_YUV444P14
Definition: pixfmt.h:409
#define AV_PIX_FMT_GBRAP10
Definition: pixfmt.h:419
int(* filter_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_yaepblur.c:52
#define AV_PIX_FMT_YUVA422P10
Definition: pixfmt.h:437
const char * desc
Definition: libsvtav1.c:79
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:71
misc image utilities
int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2613
Main libavfilter public API header.
int(* pre_calculate_row)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_yaepblur.c:51
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:168
#define AV_PIX_FMT_GBRP10
Definition: pixfmt.h:415
AVFILTER_DEFINE_CLASS(yaepblur)
#define AV_PIX_FMT_GRAY9
Definition: pixfmt.h:379
#define AV_PIX_FMT_YUV420P12
Definition: pixfmt.h:403
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
Definition: video.c:99
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
Definition: pixdesc.h:92
AVFilterFormats * ff_make_format_list(const int *fmts)
Create a list of supported formats.
Definition: formats.c:287
#define AV_PIX_FMT_GRAY10
Definition: pixfmt.h:380
#define AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
Some filters support a generic "enable" expression option that can be used to enable or disable a fil...
Definition: avfilter.h:126
const char * name
Pad name.
Definition: internal.h:60
#define AV_PIX_FMT_GRAY12
Definition: pixfmt.h:381
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1091
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:101
int planeheight[4]
Definition: vf_yaepblur.c:44
AVComponentDescriptor comp[4]
Parameters that describe how pixels are packed.
Definition: pixdesc.h:117
uint8_t
static av_cold void uninit(AVFilterContext *ctx)
Definition: vf_yaepblur.c:55
#define av_cold
Definition: attributes.h:88
AVOptions.
#define AV_PIX_FMT_YUVA420P9
Definition: pixfmt.h:433
#define AV_PIX_FMT_GBRP9
Definition: pixfmt.h:414
int height
Definition: vf_avgblur.c:61
#define height
AVFrame * dst
Definition: vf_blend.c:56
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range...
Definition: pixfmt.h:100
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:79
#define AV_PIX_FMT_YUV444P16
Definition: pixfmt.h:412
#define AV_PIX_FMT_YUV422P12
Definition: pixfmt.h:404
#define AV_PIX_FMT_YUVA420P16
Definition: pixfmt.h:441
A filter pad used for either input or output.
Definition: internal.h:54
#define src
Definition: vp8dsp.c:254
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:176
static int config_input(AVFilterLink *inlink)
Definition: vf_yaepblur.c:270
int ff_set_common_formats(AVFilterContext *ctx, AVFilterFormats *formats)
A helper for query_formats() which sets all links to the same list of formats.
Definition: formats.c:588
#define td
Definition: regdef.h:70
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
Definition: pixdesc.h:101
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:203
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
Definition: internal.h:153
const uint8_t * src
Definition: vf_bm3d.c:56
int ff_filter_process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
Generic processing of user supplied commands that are set in the same way as the filter options...
Definition: avfilter.c:885
void * priv
private data for use by the filter
Definition: avfilter.h:354
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
Definition: avfilter.h:117
#define AV_PIX_FMT_YUVA444P16
Definition: pixfmt.h:443
const char * arg
Definition: jacosubdec.c:66
#define OFFSET(x)
Definition: vf_yaepblur.c:323
#define AV_PIX_FMT_GBRAP12
Definition: pixfmt.h:420
#define AV_PIX_FMT_YUV444P10
Definition: pixfmt.h:402
int planewidth[4]
Definition: vf_yaepblur.c:43
#define AV_PIX_FMT_GBRAP16
Definition: pixfmt.h:421
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:70
#define AV_PIX_FMT_YUV422P9
Definition: pixfmt.h:397
#define AV_PIX_FMT_GBRP16
Definition: pixfmt.h:418
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
Definition: avfilter.c:800
#define AV_PIX_FMT_GRAY16
Definition: pixfmt.h:383
#define FFMIN(a, b)
Definition: common.h:96
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
int sat_linesize
Definition: vf_yaepblur.c:49
#define AV_PIX_FMT_YUVA444P12
Definition: pixfmt.h:440
#define width
uint64_t * square_sat
square summed area table
Definition: vf_yaepblur.c:48
AVFormatContext * ctx
Definition: movenc.c:48
#define s(width, name)
Definition: cbs_vp9.c:257
#define AV_PIX_FMT_YUVA444P10
Definition: pixfmt.h:438
uint64_t * sat
summed area table
Definition: vf_yaepblur.c:47
#define FILTER_SLICE(type, name)
Definition: vf_yaepblur.c:165
#define AV_PIX_FMT_YUV444P9
Definition: pixfmt.h:398
#define AV_PIX_FMT_GBRP14
Definition: pixfmt.h:417
static int query_formats(AVFilterContext *ctx)
Definition: vf_yaepblur.c:62
#define FLAGS
Definition: vf_yaepblur.c:324
static const AVFilterPad outputs[]
Definition: af_acontrast.c:203
#define AV_PIX_FMT_YUV420P16
Definition: pixfmt.h:410
static const AVFilterPad yaep_outputs[]
Definition: vf_yaepblur.c:315
static int pre_calculate_col(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_yaepblur.c:135
#define AV_PIX_FMT_YUV420P14
Definition: pixfmt.h:407
int av_frame_is_writable(AVFrame *frame)
Check if the frame data is writable.
Definition: frame.c:607
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_WB32 unsigned int_TMPL AV_WB24 unsigned int_TMPL AV_WB16 unsigned int_TMPL byte
Definition: bytestream.h:87
Used for passing data between threads.
Definition: dsddec.c:67
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:339
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:177
static const AVOption yaepblur_options[]
Definition: vf_yaepblur.c:326
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:81
AVFilter ff_vf_yaepblur
Definition: vf_yaepblur.c:338
#define AV_PIX_FMT_GRAY14
Definition: pixfmt.h:382
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several inputs
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi-0x80)*(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi-0x80)*(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t,*(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t,*(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t,*(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t,*(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31))))#define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac){}void ff_audio_convert_free(AudioConvert **ac){if(!*ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);}AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map){AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method!=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2){ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc){av_free(ac);return NULL;}return ac;}in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar){ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar?ac->channels:1;}else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;}int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in){int use_generic=1;int len=in->nb_samples;int p;if(ac->dc){av_log(ac->avr, AV_LOG_TRACE,"%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
#define AV_PIX_FMT_YUV420P10
Definition: pixfmt.h:399
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
Describe the class of an AVClass context structure.
Definition: log.h:67
Filter definition.
Definition: avfilter.h:145
const char * name
Filter name.
Definition: avfilter.h:149
#define AV_PIX_FMT_YUV440P12
Definition: pixfmt.h:405
#define AV_PIX_FMT_YUV420P9
Definition: pixfmt.h:396
AVFilterLink ** outputs
array of pointers to output links
Definition: avfilter.h:351
static enum AVPixelFormat pix_fmts[]
Definition: libkvazaar.c:300
#define AV_PIX_FMT_YUV422P14
Definition: pixfmt.h:408
#define AV_PIX_FMT_GBRP12
Definition: pixfmt.h:416
#define flags(name, subs,...)
Definition: cbs_av1.c:560
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
Definition: vf_yaepblur.c:220
AVFilterInternal * internal
An opaque struct for libavfilter internal use.
Definition: avfilter.h:379
#define AV_PIX_FMT_YUV422P10
Definition: pixfmt.h:400
#define AV_PIX_FMT_YUV444P12
Definition: pixfmt.h:406
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:322
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
int
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:66
Y , 8bpp.
Definition: pixfmt.h:74
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:215
#define AV_PIX_FMT_YUVA444P9
Definition: pixfmt.h:435
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:80
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:73
avfilter_execute_func * execute
Definition: internal.h:136
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:258
An instance of a filter.
Definition: avfilter.h:339
FILE * out
Definition: movenc.c:54
static int process_command(AVFilterContext *ctx, const char *cmd, const char *args, char *res, int res_len, int flags)
Definition: af_afftdn.c:1374
#define av_freep(p)
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:99
void av_image_copy_plane(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize, int bytewidth, int height)
Copy image plane from src to dst.
Definition: imgutils.c:373
internal API functions
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later.That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another.Frame references ownership and permissions
int depth
Number of bits in the component.
Definition: pixdesc.h:58
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
static const AVFilterPad yaep_inputs[]
Definition: vf_yaepblur.c:305
int src_linesize
Definition: vf_bm3d.c:57
#define AV_PIX_FMT_YUV422P16
Definition: pixfmt.h:411
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Definition: frame.c:671
#define AV_PIX_FMT_YUVA422P12
Definition: pixfmt.h:439
void * av_mallocz_array(size_t nmemb, size_t size)
Definition: mem.c:190
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:58
#define PRE_CALCULATE_ROW(type, name)
Definition: vf_yaepblur.c:98