FFmpeg
vf_perspective.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2002 Michael Niedermayer <michaelni@gmx.at>
3  * Copyright (c) 2013 Paul B Mahol
4  *
5  * This file is part of FFmpeg.
6  *
7  * FFmpeg is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU General Public
9  * License as published by the Free Software Foundation; either
10  * version 2 of the License, or (at your option) any later version.
11  *
12  * FFmpeg is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15  * GNU General Public License for more details.
16  *
17  * You should have received a copy of the GNU General Public License along
18  * with FFmpeg; if not, write to the Free Software
19  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20  */
21 
22 #include "libavutil/avassert.h"
23 #include "libavutil/eval.h"
24 #include "libavutil/imgutils.h"
25 #include "libavutil/pixdesc.h"
26 #include "libavutil/opt.h"
27 #include "avfilter.h"
28 #include "formats.h"
29 #include "internal.h"
30 #include "video.h"
31 
32 #define SUB_PIXEL_BITS 8
33 #define SUB_PIXELS (1 << SUB_PIXEL_BITS)
34 #define COEFF_BITS 11
35 
36 #define LINEAR 0
37 #define CUBIC 1
38 
39 typedef struct PerspectiveContext {
40  const AVClass *class;
41  char *expr_str[4][2];
42  double ref[4][2];
43  int32_t (*pv)[2];
46  int linesize[4];
47  int height[4];
48  int hsub, vsub;
49  int nb_planes;
50  int sense;
51  int eval_mode;
52 
54  void *arg, int job, int nb_jobs);
56 
57 #define OFFSET(x) offsetof(PerspectiveContext, x)
58 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
59 
61  PERSPECTIVE_SENSE_SOURCE = 0, ///< coordinates give locations in source of corners of destination.
62  PERSPECTIVE_SENSE_DESTINATION = 1, ///< coordinates give locations in destination of corners of source.
63 };
64 
65 enum EvalMode {
69 };
70 
71 static const AVOption perspective_options[] = {
72  { "x0", "set top left x coordinate", OFFSET(expr_str[0][0]), AV_OPT_TYPE_STRING, {.str="0"}, 0, 0, FLAGS },
73  { "y0", "set top left y coordinate", OFFSET(expr_str[0][1]), AV_OPT_TYPE_STRING, {.str="0"}, 0, 0, FLAGS },
74  { "x1", "set top right x coordinate", OFFSET(expr_str[1][0]), AV_OPT_TYPE_STRING, {.str="W"}, 0, 0, FLAGS },
75  { "y1", "set top right y coordinate", OFFSET(expr_str[1][1]), AV_OPT_TYPE_STRING, {.str="0"}, 0, 0, FLAGS },
76  { "x2", "set bottom left x coordinate", OFFSET(expr_str[2][0]), AV_OPT_TYPE_STRING, {.str="0"}, 0, 0, FLAGS },
77  { "y2", "set bottom left y coordinate", OFFSET(expr_str[2][1]), AV_OPT_TYPE_STRING, {.str="H"}, 0, 0, FLAGS },
78  { "x3", "set bottom right x coordinate", OFFSET(expr_str[3][0]), AV_OPT_TYPE_STRING, {.str="W"}, 0, 0, FLAGS },
79  { "y3", "set bottom right y coordinate", OFFSET(expr_str[3][1]), AV_OPT_TYPE_STRING, {.str="H"}, 0, 0, FLAGS },
80  { "interpolation", "set interpolation", OFFSET(interpolation), AV_OPT_TYPE_INT, {.i64=LINEAR}, 0, 1, FLAGS, "interpolation" },
81  { "linear", "", 0, AV_OPT_TYPE_CONST, {.i64=LINEAR}, 0, 0, FLAGS, "interpolation" },
82  { "cubic", "", 0, AV_OPT_TYPE_CONST, {.i64=CUBIC}, 0, 0, FLAGS, "interpolation" },
83  { "sense", "specify the sense of the coordinates", OFFSET(sense), AV_OPT_TYPE_INT, {.i64=PERSPECTIVE_SENSE_SOURCE}, 0, 1, FLAGS, "sense"},
84  { "source", "specify locations in source to send to corners in destination",
85  0, AV_OPT_TYPE_CONST, {.i64=PERSPECTIVE_SENSE_SOURCE}, 0, 0, FLAGS, "sense"},
86  { "destination", "specify locations in destination to send corners of source",
87  0, AV_OPT_TYPE_CONST, {.i64=PERSPECTIVE_SENSE_DESTINATION}, 0, 0, FLAGS, "sense"},
88  { "eval", "specify when to evaluate expressions", OFFSET(eval_mode), AV_OPT_TYPE_INT, {.i64 = EVAL_MODE_INIT}, 0, EVAL_MODE_NB-1, FLAGS, "eval" },
89  { "init", "eval expressions once during initialization", 0, AV_OPT_TYPE_CONST, {.i64=EVAL_MODE_INIT}, .flags = FLAGS, .unit = "eval" },
90  { "frame", "eval expressions per-frame", 0, AV_OPT_TYPE_CONST, {.i64=EVAL_MODE_FRAME}, .flags = FLAGS, .unit = "eval" },
91 
92  { NULL }
93 };
94 
95 AVFILTER_DEFINE_CLASS(perspective);
96 
97 static const enum AVPixelFormat pix_fmts[] = {
102 };
103 
104 static inline double get_coeff(double d)
105 {
106  double coeff, A = -0.60;
107 
108  d = fabs(d);
109 
110  if (d < 1.0)
111  coeff = (1.0 - (A + 3.0) * d * d + (A + 2.0) * d * d * d);
112  else if (d < 2.0)
113  coeff = (-4.0 * A + 8.0 * A * d - 5.0 * A * d * d + A * d * d * d);
114  else
115  coeff = 0.0;
116 
117  return coeff;
118 }
119 
120 static const char *const var_names[] = { "W", "H", "in", "on", NULL };
122 
124 {
125  PerspectiveContext *s = ctx->priv;
126  AVFilterLink *outlink = ctx->outputs[0];
127  double (*ref)[2] = s->ref;
128 
129  double values[VAR_VARS_NB] = { [VAR_W] = inlink->w, [VAR_H] = inlink->h,
130  [VAR_IN] = inlink->frame_count_out + 1,
131  [VAR_ON] = outlink->frame_count_in + 1 };
132  const int h = values[VAR_H];
133  const int w = values[VAR_W];
134  double x0, x1, x2, x3, x4, x5, x6, x7, x8, q;
135  double t0, t1, t2, t3;
136  int x, y, i, j, ret;
137 
138  for (i = 0; i < 4; i++) {
139  for (j = 0; j < 2; j++) {
140  if (!s->expr_str[i][j])
141  return AVERROR(EINVAL);
142  ret = av_expr_parse_and_eval(&s->ref[i][j], s->expr_str[i][j],
143  var_names, &values[0],
144  NULL, NULL, NULL, NULL,
145  0, 0, ctx);
146  if (ret < 0)
147  return ret;
148  }
149  }
150 
151  switch (s->sense) {
153  x6 = ((ref[0][0] - ref[1][0] - ref[2][0] + ref[3][0]) *
154  (ref[2][1] - ref[3][1]) -
155  ( ref[0][1] - ref[1][1] - ref[2][1] + ref[3][1]) *
156  (ref[2][0] - ref[3][0])) * h;
157  x7 = ((ref[0][1] - ref[1][1] - ref[2][1] + ref[3][1]) *
158  (ref[1][0] - ref[3][0]) -
159  ( ref[0][0] - ref[1][0] - ref[2][0] + ref[3][0]) *
160  (ref[1][1] - ref[3][1])) * w;
161  q = ( ref[1][0] - ref[3][0]) * (ref[2][1] - ref[3][1]) -
162  ( ref[2][0] - ref[3][0]) * (ref[1][1] - ref[3][1]);
163 
164  x0 = q * (ref[1][0] - ref[0][0]) * h + x6 * ref[1][0];
165  x1 = q * (ref[2][0] - ref[0][0]) * w + x7 * ref[2][0];
166  x2 = q * ref[0][0] * w * h;
167  x3 = q * (ref[1][1] - ref[0][1]) * h + x6 * ref[1][1];
168  x4 = q * (ref[2][1] - ref[0][1]) * w + x7 * ref[2][1];
169  x5 = q * ref[0][1] * w * h;
170  x8 = q * w * h;
171  break;
173  t0 = ref[0][0] * (ref[3][1] - ref[1][1]) +
174  ref[1][0] * (ref[0][1] - ref[3][1]) +
175  ref[3][0] * (ref[1][1] - ref[0][1]);
176  t1 = ref[1][0] * (ref[2][1] - ref[3][1]) +
177  ref[2][0] * (ref[3][1] - ref[1][1]) +
178  ref[3][0] * (ref[1][1] - ref[2][1]);
179  t2 = ref[0][0] * (ref[3][1] - ref[2][1]) +
180  ref[2][0] * (ref[0][1] - ref[3][1]) +
181  ref[3][0] * (ref[2][1] - ref[0][1]);
182  t3 = ref[0][0] * (ref[1][1] - ref[2][1]) +
183  ref[1][0] * (ref[2][1] - ref[0][1]) +
184  ref[2][0] * (ref[0][1] - ref[1][1]);
185 
186  x0 = t0 * t1 * w * (ref[2][1] - ref[0][1]);
187  x1 = t0 * t1 * w * (ref[0][0] - ref[2][0]);
188  x2 = t0 * t1 * w * (ref[0][1] * ref[2][0] - ref[0][0] * ref[2][1]);
189  x3 = t1 * t2 * h * (ref[1][1] - ref[0][1]);
190  x4 = t1 * t2 * h * (ref[0][0] - ref[1][0]);
191  x5 = t1 * t2 * h * (ref[0][1] * ref[1][0] - ref[0][0] * ref[1][1]);
192  x6 = t1 * t2 * (ref[1][1] - ref[0][1]) +
193  t0 * t3 * (ref[2][1] - ref[3][1]);
194  x7 = t1 * t2 * (ref[0][0] - ref[1][0]) +
195  t0 * t3 * (ref[3][0] - ref[2][0]);
196  x8 = t1 * t2 * (ref[0][1] * ref[1][0] - ref[0][0] * ref[1][1]) +
197  t0 * t3 * (ref[2][0] * ref[3][1] - ref[2][1] * ref[3][0]);
198  break;
199  default:
200  av_assert0(0);
201  }
202 
203  for (y = 0; y < h; y++){
204  for (x = 0; x < w; x++){
205  int u, v;
206 
207  u = lrint(SUB_PIXELS * (x0 * x + x1 * y + x2) /
208  (x6 * x + x7 * y + x8));
209  v = lrint(SUB_PIXELS * (x3 * x + x4 * y + x5) /
210  (x6 * x + x7 * y + x8));
211 
212  s->pv[x + y * w][0] = u;
213  s->pv[x + y * w][1] = v;
214  }
215  }
216 
217  return 0;
218 }
219 
221 {
222  AVFilterContext *ctx = inlink->dst;
223  PerspectiveContext *s = ctx->priv;
225  int h = inlink->h;
226  int w = inlink->w;
227  int i, j, ret;
228  s->hsub = desc->log2_chroma_w;
229  s->vsub = desc->log2_chroma_h;
230  s->nb_planes = av_pix_fmt_count_planes(inlink->format);
231  if ((ret = av_image_fill_linesizes(s->linesize, inlink->format, inlink->w)) < 0)
232  return ret;
233 
234  s->height[1] = s->height[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
235  s->height[0] = s->height[3] = inlink->h;
236 
237  s->pv = av_realloc_f(s->pv, w * h, 2 * sizeof(*s->pv));
238  if (!s->pv)
239  return AVERROR(ENOMEM);
240 
241  if (s->eval_mode == EVAL_MODE_INIT) {
242  if ((ret = calc_persp_luts(ctx, inlink)) < 0) {
243  return ret;
244  }
245  }
246 
247  for (i = 0; i < SUB_PIXELS; i++){
248  double d = i / (double)SUB_PIXELS;
249  double temp[4];
250  double sum = 0;
251 
252  for (j = 0; j < 4; j++)
253  temp[j] = get_coeff(j - d - 1);
254 
255  for (j = 0; j < 4; j++)
256  sum += temp[j];
257 
258  for (j = 0; j < 4; j++)
259  s->coeff[i][j] = lrint((1 << COEFF_BITS) * temp[j] / sum);
260  }
261 
262  return 0;
263 }
264 
265 typedef struct ThreadData {
266  uint8_t *dst;
267  int dst_linesize;
268  uint8_t *src;
269  int src_linesize;
270  int w, h;
271  int hsub, vsub;
272 } ThreadData;
273 
275  int job, int nb_jobs)
276 {
277  PerspectiveContext *s = ctx->priv;
278  ThreadData *td = arg;
279  uint8_t *dst = td->dst;
280  int dst_linesize = td->dst_linesize;
281  uint8_t *src = td->src;
282  int src_linesize = td->src_linesize;
283  int w = td->w;
284  int h = td->h;
285  int hsub = td->hsub;
286  int vsub = td->vsub;
287  int start = (h * job) / nb_jobs;
288  int end = (h * (job+1)) / nb_jobs;
289  const int linesize = s->linesize[0];
290  int x, y;
291 
292  for (y = start; y < end; y++) {
293  int sy = y << vsub;
294  for (x = 0; x < w; x++) {
295  int u, v, subU, subV, sum, sx;
296 
297  sx = x << hsub;
298  u = s->pv[sx + sy * linesize][0] >> hsub;
299  v = s->pv[sx + sy * linesize][1] >> vsub;
300  subU = u & (SUB_PIXELS - 1);
301  subV = v & (SUB_PIXELS - 1);
302  u >>= SUB_PIXEL_BITS;
303  v >>= SUB_PIXEL_BITS;
304 
305  if (u > 0 && v > 0 && u < w - 2 && v < h - 2){
306  const int index = u + v*src_linesize;
307  const int a = s->coeff[subU][0];
308  const int b = s->coeff[subU][1];
309  const int c = s->coeff[subU][2];
310  const int d = s->coeff[subU][3];
311 
312  sum = s->coeff[subV][0] * (a * src[index - 1 - src_linesize] + b * src[index - 0 - src_linesize] +
313  c * src[index + 1 - src_linesize] + d * src[index + 2 - src_linesize]) +
314  s->coeff[subV][1] * (a * src[index - 1 ] + b * src[index - 0 ] +
315  c * src[index + 1 ] + d * src[index + 2 ]) +
316  s->coeff[subV][2] * (a * src[index - 1 + src_linesize] + b * src[index - 0 + src_linesize] +
317  c * src[index + 1 + src_linesize] + d * src[index + 2 + src_linesize]) +
318  s->coeff[subV][3] * (a * src[index - 1 + 2 * src_linesize] + b * src[index - 0 + 2 * src_linesize] +
319  c * src[index + 1 + 2 * src_linesize] + d * src[index + 2 + 2 * src_linesize]);
320  } else {
321  int dx, dy;
322 
323  sum = 0;
324 
325  for (dy = 0; dy < 4; dy++) {
326  int iy = v + dy - 1;
327 
328  if (iy < 0)
329  iy = 0;
330  else if (iy >= h)
331  iy = h-1;
332  for (dx = 0; dx < 4; dx++) {
333  int ix = u + dx - 1;
334 
335  if (ix < 0)
336  ix = 0;
337  else if (ix >= w)
338  ix = w - 1;
339 
340  sum += s->coeff[subU][dx] * s->coeff[subV][dy] * src[ ix + iy * src_linesize];
341  }
342  }
343  }
344 
345  sum = (sum + (1<<(COEFF_BITS * 2 - 1))) >> (COEFF_BITS * 2);
346  sum = av_clip_uint8(sum);
347  dst[x + y * dst_linesize] = sum;
348  }
349  }
350  return 0;
351 }
352 
354  int job, int nb_jobs)
355 {
356  PerspectiveContext *s = ctx->priv;
357  ThreadData *td = arg;
358  uint8_t *dst = td->dst;
359  int dst_linesize = td->dst_linesize;
360  uint8_t *src = td->src;
361  int src_linesize = td->src_linesize;
362  int w = td->w;
363  int h = td->h;
364  int hsub = td->hsub;
365  int vsub = td->vsub;
366  int start = (h * job) / nb_jobs;
367  int end = (h * (job+1)) / nb_jobs;
368  const int linesize = s->linesize[0];
369  int x, y;
370 
371  for (y = start; y < end; y++){
372  int sy = y << vsub;
373  for (x = 0; x < w; x++){
374  int u, v, subU, subV, sum, sx, index, subUI, subVI;
375 
376  sx = x << hsub;
377  u = s->pv[sx + sy * linesize][0] >> hsub;
378  v = s->pv[sx + sy * linesize][1] >> vsub;
379  subU = u & (SUB_PIXELS - 1);
380  subV = v & (SUB_PIXELS - 1);
381  u >>= SUB_PIXEL_BITS;
382  v >>= SUB_PIXEL_BITS;
383 
384  index = u + v * src_linesize;
385  subUI = SUB_PIXELS - subU;
386  subVI = SUB_PIXELS - subV;
387 
388  if ((unsigned)u < (unsigned)(w - 1)){
389  if((unsigned)v < (unsigned)(h - 1)){
390  sum = subVI * (subUI * src[index] + subU * src[index + 1]) +
391  subV * (subUI * src[index + src_linesize] + subU * src[index + src_linesize + 1]);
392  sum = (sum + (1 << (SUB_PIXEL_BITS * 2 - 1)))>> (SUB_PIXEL_BITS * 2);
393  } else {
394  if (v < 0)
395  v = 0;
396  else
397  v = h - 1;
398  index = u + v * src_linesize;
399  sum = subUI * src[index] + subU * src[index + 1];
400  sum = (sum + (1 << (SUB_PIXEL_BITS - 1))) >> SUB_PIXEL_BITS;
401  }
402  } else {
403  if (u < 0)
404  u = 0;
405  else
406  u = w - 1;
407  if ((unsigned)v < (unsigned)(h - 1)){
408  index = u + v * src_linesize;
409  sum = subVI * src[index] + subV * src[index + src_linesize];
410  sum = (sum + (1 << (SUB_PIXEL_BITS - 1))) >> SUB_PIXEL_BITS;
411  } else {
412  if (v < 0)
413  v = 0;
414  else
415  v = h - 1;
416  index = u + v * src_linesize;
417  sum = src[index];
418  }
419  }
420 
421  sum = av_clip_uint8(sum);
422  dst[x + y * dst_linesize] = sum;
423  }
424  }
425  return 0;
426 }
427 
429 {
430  PerspectiveContext *s = ctx->priv;
431 
432  switch (s->interpolation) {
433  case LINEAR: s->perspective = resample_linear; break;
434  case CUBIC: s->perspective = resample_cubic; break;
435  }
436 
437  return 0;
438 }
439 
441 {
442  AVFilterContext *ctx = inlink->dst;
443  AVFilterLink *outlink = ctx->outputs[0];
444  PerspectiveContext *s = ctx->priv;
445  AVFrame *out;
446  int plane;
447  int ret;
448 
449  out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
450  if (!out) {
452  return AVERROR(ENOMEM);
453  }
455 
456  if (s->eval_mode == EVAL_MODE_FRAME) {
457  if ((ret = calc_persp_luts(ctx, inlink)) < 0) {
458  av_frame_free(&out);
459  return ret;
460  }
461  }
462 
463  for (plane = 0; plane < s->nb_planes; plane++) {
464  int hsub = plane == 1 || plane == 2 ? s->hsub : 0;
465  int vsub = plane == 1 || plane == 2 ? s->vsub : 0;
466  ThreadData td = {.dst = out->data[plane],
467  .dst_linesize = out->linesize[plane],
468  .src = frame->data[plane],
469  .src_linesize = frame->linesize[plane],
470  .w = s->linesize[plane],
471  .h = s->height[plane],
472  .hsub = hsub,
473  .vsub = vsub };
474  ff_filter_execute(ctx, s->perspective, &td, NULL,
476  }
477 
479  return ff_filter_frame(outlink, out);
480 }
481 
483 {
484  PerspectiveContext *s = ctx->priv;
485 
486  av_freep(&s->pv);
487 }
488 
489 static const AVFilterPad perspective_inputs[] = {
490  {
491  .name = "default",
492  .type = AVMEDIA_TYPE_VIDEO,
493  .filter_frame = filter_frame,
494  .config_props = config_input,
495  },
496 };
497 
499  {
500  .name = "default",
501  .type = AVMEDIA_TYPE_VIDEO,
502  },
503 };
504 
506  .name = "perspective",
507  .description = NULL_IF_CONFIG_SMALL("Correct the perspective of video."),
508  .priv_size = sizeof(PerspectiveContext),
509  .init = init,
510  .uninit = uninit,
514  .priv_class = &perspective_class,
516 };
get_coeff
static double get_coeff(double d)
Definition: vf_perspective.c:104
ff_get_video_buffer
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
Definition: video.c:98
PerspectiveContext::sense
int sense
Definition: vf_perspective.c:50
ThreadData::hsub
int hsub
Definition: vf_perspective.c:271
VAR_VARS_NB
@ VAR_VARS_NB
Definition: vf_perspective.c:121
PerspectiveContext::coeff
int32_t coeff[SUB_PIXELS][4]
Definition: vf_perspective.c:44
td
#define td
Definition: regdef.h:70
AVPixelFormat
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
pix_fmts
static enum AVPixelFormat pix_fmts[]
Definition: vf_perspective.c:97
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
opt.h
out
FILE * out
Definition: movenc.c:54
u
#define u(width, name, range_min, range_max)
Definition: cbs_h2645.c:264
PerspectiveContext::height
int height[4]
Definition: vf_perspective.c:47
PerspectiveContext::expr_str
char * expr_str[4][2]
Definition: vf_perspective.c:41
ff_filter_frame
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1018
av_pix_fmt_desc_get
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2660
FILTER_PIXFMTS_ARRAY
#define FILTER_PIXFMTS_ARRAY(array)
Definition: internal.h:171
inlink
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
Definition: filter_design.txt:212
PerspectiveContext::interpolation
int interpolation
Definition: vf_perspective.c:45
av_frame_free
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:109
PerspectiveContext::perspective
int(* perspective)(AVFilterContext *ctx, void *arg, int job, int nb_jobs)
Definition: vf_perspective.c:53
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:317
pixdesc.h
index
fg index
Definition: ffmpeg_filter.c:167
EVAL_MODE_FRAME
@ EVAL_MODE_FRAME
Definition: vf_perspective.c:67
w
uint8_t w
Definition: llviddspenc.c:38
calc_persp_luts
static int calc_persp_luts(AVFilterContext *ctx, AVFilterLink *inlink)
Definition: vf_perspective.c:123
AVOption
AVOption.
Definition: opt.h:247
t0
#define t0
Definition: regdef.h:28
b
#define b
Definition: input.c:40
PerspectiveContext::nb_planes
int nb_planes
Definition: vf_perspective.c:49
PERSPECTIVESense
PERSPECTIVESense
Definition: vf_perspective.c:60
t1
#define t1
Definition: regdef.h:29
ThreadData::w
int w
Definition: vf_blend.c:86
AV_PIX_FMT_YUV440P
@ AV_PIX_FMT_YUV440P
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:99
AVFilter::name
const char * name
Filter name.
Definition: avfilter.h:169
VAR_IN
@ VAR_IN
Definition: vf_perspective.c:121
video.h
PERSPECTIVE_SENSE_SOURCE
@ PERSPECTIVE_SENSE_SOURCE
coordinates give locations in source of corners of destination.
Definition: vf_perspective.c:61
hsub
static void hsub(htype *dst, const htype *src, int bins)
Definition: vf_median.c:74
VAR_W
@ VAR_W
Definition: vf_perspective.c:121
formats.h
A
#define A(x)
Definition: vp56_arith.h:28
VAR_ON
@ VAR_ON
Definition: vf_perspective.c:121
av_pix_fmt_count_planes
int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2700
AV_PIX_FMT_GBRAP
@ AV_PIX_FMT_GBRAP
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:205
PerspectiveContext::ref
double ref[4][2]
Definition: vf_perspective.c:42
COEFF_BITS
#define COEFF_BITS
Definition: vf_perspective.c:34
AVFilterPad
A filter pad used for either input or output.
Definition: internal.h:50
AV_PIX_FMT_YUVJ411P
@ AV_PIX_FMT_YUVJ411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:248
EVAL_MODE_NB
@ EVAL_MODE_NB
Definition: vf_perspective.c:68
avassert.h
lrint
#define lrint
Definition: tablegen.h:53
av_cold
#define av_cold
Definition: attributes.h:90
init
static av_cold int init(AVFilterContext *ctx)
Definition: vf_perspective.c:428
SUB_PIXEL_BITS
#define SUB_PIXEL_BITS
Definition: vf_perspective.c:32
AV_PIX_FMT_YUVJ422P
@ AV_PIX_FMT_YUVJ422P
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:79
resample_linear
static int resample_linear(AVFilterContext *ctx, void *arg, int job, int nb_jobs)
Definition: vf_perspective.c:353
av_image_fill_linesizes
int av_image_fill_linesizes(int linesizes[4], enum AVPixelFormat pix_fmt, int width)
Fill plane linesizes for an image with pixel format pix_fmt and width width.
Definition: imgutils.c:89
s
#define s(width, name)
Definition: cbs_vp9.c:257
AV_PIX_FMT_YUVA420P
@ AV_PIX_FMT_YUVA420P
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:101
AV_CEIL_RSHIFT
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:51
PerspectiveContext::linesize
int linesize[4]
Definition: vf_perspective.c:46
LINEAR
#define LINEAR
Definition: vf_perspective.c:36
ThreadData::src
uint8_t * src
Definition: vf_perspective.c:268
av_assert0
#define av_assert0(cond)
assert() equivalent, that is always enabled.
Definition: avassert.h:37
ctx
AVFormatContext * ctx
Definition: movenc.c:48
AV_PIX_FMT_YUV420P
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:66
FILTER_INPUTS
#define FILTER_INPUTS(array)
Definition: internal.h:191
ThreadData::h
int h
Definition: vf_blend.c:86
AV_PIX_FMT_YUVJ444P
@ AV_PIX_FMT_YUVJ444P
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:80
arg
const char * arg
Definition: jacosubdec.c:67
av_realloc_f
#define av_realloc_f(p, o, n)
Definition: tableprint_vlc.h:33
perspective_options
static const AVOption perspective_options[]
Definition: vf_perspective.c:71
perspective_inputs
static const AVFilterPad perspective_inputs[]
Definition: vf_perspective.c:489
AVClass
Describe the class of an AVClass context structure.
Definition: log.h:66
fabs
static __device__ float fabs(float a)
Definition: cuda_runtime.h:182
NULL
#define NULL
Definition: coverity.c:32
av_frame_copy_props
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Definition: frame.c:537
AV_PIX_FMT_YUVJ420P
@ AV_PIX_FMT_YUVJ420P
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
src
#define src
Definition: vp8dsp.c:255
AV_PIX_FMT_GRAY8
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
Definition: pixfmt.h:74
c
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
Definition: undefined.txt:32
eval.h
NULL_IF_CONFIG_SMALL
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
Definition: internal.h:117
av_expr_parse_and_eval
int av_expr_parse_and_eval(double *d, const char *s, const char *const *const_names, const double *const_values, const char *const *func1_names, double(*const *funcs1)(void *, double), const char *const *func2_names, double(*const *funcs2)(void *, double, double), void *opaque, int log_offset, void *log_ctx)
Parse and evaluate an expression.
Definition: eval.c:776
OFFSET
#define OFFSET(x)
Definition: vf_perspective.c:57
ff_vf_perspective
const AVFilter ff_vf_perspective
Definition: vf_perspective.c:505
perspective_outputs
static const AVFilterPad perspective_outputs[]
Definition: vf_perspective.c:498
PERSPECTIVE_SENSE_DESTINATION
@ PERSPECTIVE_SENSE_DESTINATION
coordinates give locations in destination of corners of source.
Definition: vf_perspective.c:62
AVFILTER_DEFINE_CLASS
AVFILTER_DEFINE_CLASS(perspective)
a
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
Definition: undefined.txt:41
AV_PIX_FMT_YUVA444P
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:167
interpolation
static int interpolation(DeclickChannel *c, const double *src, int ar_order, double *acoefficients, int *index, int nb_errors, double *auxiliary, double *interpolated)
Definition: af_adeclick.c:336
internal.h
AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
#define AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
Some filters support a generic "enable" expression option that can be used to enable or disable a fil...
Definition: avfilter.h:146
i
#define i(width, name, range_min, range_max)
Definition: cbs_h2645.c:271
t3
#define t3
Definition: regdef.h:31
ThreadData::dst
uint8_t * dst
Definition: vf_perspective.c:266
ff_filter_get_nb_threads
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
Definition: avfilter.c:803
ThreadData
Used for passing data between threads.
Definition: dsddec.c:67
EvalMode
EvalMode
Definition: af_volume.h:39
FFMIN
#define FFMIN(a, b)
Definition: macros.h:49
config_input
static int config_input(AVFilterLink *inlink)
Definition: vf_perspective.c:220
AV_PIX_FMT_YUVJ440P
@ AV_PIX_FMT_YUVJ440P
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range
Definition: pixfmt.h:100
AVFilterPad::name
const char * name
Pad name.
Definition: internal.h:56
uninit
static av_cold void uninit(AVFilterContext *ctx)
Definition: vf_perspective.c:482
AVFilter
Filter definition.
Definition: avfilter.h:165
ret
ret
Definition: filter_design.txt:187
frame
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
Definition: filter_design.txt:264
PerspectiveContext
Definition: vf_perspective.c:39
ThreadData::vsub
int vsub
Definition: vf_perspective.c:271
EVAL_MODE_INIT
@ EVAL_MODE_INIT
Definition: vf_perspective.c:66
PerspectiveContext::vsub
int vsub
Definition: vf_perspective.c:48
ThreadData::src_linesize
int src_linesize
Definition: vf_bm3d.c:56
resample_cubic
static int resample_cubic(AVFilterContext *ctx, void *arg, int job, int nb_jobs)
Definition: vf_perspective.c:274
t2
#define t2
Definition: regdef.h:30
PerspectiveContext::hsub
int hsub
Definition: vf_perspective.c:48
AV_PIX_FMT_NONE
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:65
AV_OPT_TYPE_INT
@ AV_OPT_TYPE_INT
Definition: opt.h:224
avfilter.h
var_names
static const char *const var_names[]
Definition: vf_perspective.c:120
values
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return values
Definition: filter_design.txt:263
ref
static int ref[MAX_W *MAX_W]
Definition: jpeg2000dwt.c:107
temp
else temp
Definition: vf_mcdeint.c:248
SUB_PIXELS
#define SUB_PIXELS
Definition: vf_perspective.c:33
av_clip_uint8
#define av_clip_uint8
Definition: common.h:102
AV_PIX_FMT_YUV444P
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:71
AVFilterContext
An instance of a filter.
Definition: avfilter.h:402
AV_PIX_FMT_GBRP
@ AV_PIX_FMT_GBRP
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:158
AVFILTER_FLAG_SLICE_THREADS
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
Definition: avfilter.h:121
desc
const char * desc
Definition: libsvtav1.c:79
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
AV_PIX_FMT_YUV422P
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:70
ThreadData::dst_linesize
int dst_linesize
Definition: vf_colorlevels.c:90
AVPixFmtDescriptor
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:69
FILTER_OUTPUTS
#define FILTER_OUTPUTS(array)
Definition: internal.h:192
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:35
AV_PIX_FMT_YUV411P
@ AV_PIX_FMT_YUV411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:73
PerspectiveContext::eval_mode
int eval_mode
Definition: vf_perspective.c:51
d
d
Definition: ffmpeg_filter.c:153
int32_t
int32_t
Definition: audioconvert.c:56
imgutils.h
AV_PIX_FMT_YUV410P
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
coeff
static const double coeff[2][5]
Definition: vf_owdenoise.c:78
filter_frame
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
Definition: vf_perspective.c:440
h
h
Definition: vp9dsp_template.c:2038
AV_OPT_TYPE_STRING
@ AV_OPT_TYPE_STRING
Definition: opt.h:228
ff_filter_execute
static av_always_inline int ff_filter_execute(AVFilterContext *ctx, avfilter_action_func *func, void *arg, int *ret, int nb_jobs)
Definition: internal.h:143
int
int
Definition: ffmpeg_filter.c:153
PerspectiveContext::pv
int32_t(* pv)[2]
Definition: vf_perspective.c:43
FLAGS
#define FLAGS
Definition: vf_perspective.c:58
AV_OPT_TYPE_CONST
@ AV_OPT_TYPE_CONST
Definition: opt.h:233
AV_PIX_FMT_YUVA422P
@ AV_PIX_FMT_YUVA422P
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:166
VAR_H
@ VAR_H
Definition: vf_perspective.c:121
CUBIC
#define CUBIC
Definition: vf_perspective.c:37