38 #define MV_P_FOR (1<<0)
39 #define MV_B_FOR (1<<1)
40 #define MV_B_BACK (1<<2)
41 #define MV_TYPE_FOR (1<<0)
42 #define MV_TYPE_BACK (1<<1)
43 #define FRAME_TYPE_I (1<<0)
44 #define FRAME_TYPE_P (1<<1)
45 #define FRAME_TYPE_B (1<<2)
56 #define OFFSET(x) offsetof(CodecViewContext, x)
57 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
58 #define CONST(name, help, val, unit) { name, help, 0, AV_OPT_TYPE_CONST, {.i64=val}, 0, 0, FLAGS, unit }
62 CONST(
"pf",
"forward predicted MVs of P-frames",
MV_P_FOR,
"mv"),
63 CONST(
"bf",
"forward predicted MVs of B-frames",
MV_B_FOR,
"mv"),
64 CONST(
"bb",
"backward predicted MVs of B-frames",
MV_B_BACK,
"mv"),
70 {
"frame_type",
"set frame types to visualize motion vectors of",
OFFSET(frame_type),
AV_OPT_TYPE_FLAGS, {.i64=0}, 0, INT_MAX,
FLAGS,
"frame_type" },
71 {
"ft",
"set frame types to visualize motion vectors of",
OFFSET(frame_type),
AV_OPT_TYPE_FLAGS, {.i64=0}, 0, INT_MAX,
FLAGS,
"frame_type" },
91 static int clip_line(
int *sx,
int *sy,
int *ex,
int *ey,
int maxx)
99 *sy = *ey + (*sy - *ey) * (int64_t)*ex / (*ex - *sx);
106 *ey = *sy + (*ey - *sy) * (int64_t)(maxx - *sx) / (*ex - *sx);
124 if (
clip_line(&sx, &sy, &ex, &ey, w - 1))
126 if (
clip_line(&sy, &sx, &ey, &ex, h - 1))
129 sx = av_clip(sx, 0, w - 1);
130 sy = av_clip(sy, 0, h - 1);
131 ex = av_clip(ex, 0, w - 1);
132 ey = av_clip(ey, 0, h - 1);
134 buf[sy * stride + sx] +=
color;
143 f = ((ey - sy) << 16) / ex;
144 for (x = 0; x <= ex; x++) {
146 fr = (x * f) & 0xFFFF;
147 buf[ y * stride + x] += (color * (0x10000 - fr)) >> 16;
148 if(fr) buf[(y + 1) * stride + x] += (color * fr ) >> 16;
158 f = ((ex - sx) << 16) / ey;
161 for(y= 0; y <= ey; y++){
164 buf[y * stride + x ] += (color * (0x10000 - fr)) >> 16;
165 if(fr) buf[y * stride + x + 1] += (color * fr ) >> 16;
178 int ey,
int w,
int h,
int stride,
int color,
int tail,
int direction)
187 sx = av_clip(sx, -100, w + 100);
188 sy = av_clip(sy, -100, h + 100);
189 ex = av_clip(ex, -100, w + 100);
190 ey = av_clip(ey, -100, h + 100);
195 if (dx * dx + dy * dy > 3 * 3) {
198 int length = sqrt((rx * rx + ry * ry) << 8);
209 draw_line(buf, sx, sy, sx + rx, sy + ry, w, h, stride, color);
210 draw_line(buf, sx, sy, sx - ry, sy + rx, w, h, stride, color);
212 draw_line(buf, sx, sy, ex, ey, w, h, stride, color);
222 int qstride, qp_type;
234 for (y = 0; y <
h; y++) {
235 for (x = 0; x < w; x++) {
236 const int qp =
ff_norm_qscale(qp_table[(y >> 3) * qstride + (x >> 3)], qp_type) * 128/31;
254 for (i = 0; i < sd->
size /
sizeof(*mvs); i++) {
256 const int direction = mv->
source > 0;
263 is_iframe && is_fp || is_iframe && is_bp ||
264 is_pframe && is_fp ||
265 is_bframe && is_fp || is_bframe && is_bp)
318 .
inputs = codecview_inputs,
320 .priv_class = &codecview_class,
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
This structure describes decoded (raw) audio or video data.
int16_t src_x
Absolute source position.
Main libavfilter public API header.
static const AVFilterPad codecview_outputs[]
static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
static const AVOption codecview_options[]
int8_t * av_frame_get_qp_table(AVFrame *f, int *stride, int *type)
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
AVFrameSideData * av_frame_get_side_data(const AVFrame *frame, enum AVFrameSideDataType type)
#define AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC
Some filters support a generic "enable" expression option that can be used to enable or disable a fil...
const char * name
Pad name.
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
static const uint32_t color[16+AV_CLASS_CATEGORY_NB]
int16_t dst_x
Absolute destination position.
Structure to hold side data for an AVFrame.
int32_t source
Where the current macroblock comes from; negative value when it comes from the past, positive value when it comes from the future.
static void draw_arrow(uint8_t *buf, int sx, int sy, int ex, int ey, int w, int h, int stride, int color, int tail, int direction)
Draw an arrow from (ex, ey) -> (sx, sy).
#define ROUNDED_DIV(a, b)
A filter pad used for either input or output.
static void draw_line(uint8_t *buf, int sx, int sy, int ex, int ey, int w, int h, int stride, int color)
Draw a line from (ex, ey) -> (sx, sy).
A link between two filters.
#define CONST(name, help, val, unit)
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
void * priv
private data for use by the filter
AVFILTER_DEFINE_CLASS(codecview)
enum AVPictureType pict_type
Picture type of the frame.
Motion vectors exported by some codecs (on demand through the export_mvs flag set in the libavcodec A...
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
static const AVFilterPad outputs[]
int format
agreed upon media format
static const int8_t mv[256][2]
static const AVFilterPad inputs[]
static int query_formats(AVFilterContext *ctx)
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
static int clip_line(int *sx, int *sy, int *ex, int *ey, int maxx)
Describe the class of an AVClass context structure.
const char * name
Filter name.
AVFilterLink ** outputs
array of pointers to output links
static int ff_norm_qscale(int qscale, int type)
Normalize the qscale factor FIXME the H264 qscale is a log based scale, mpeg1/2 is not...
static enum AVPixelFormat pix_fmts[]
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
static int config_input(AVFilterLink *inlink)
static const AVFilterPad codecview_inputs[]
GLint GLenum GLboolean GLsizei stride
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
AVFilterContext * dst
dest filter
#define FFSWAP(type, a, b)
AVPixelFormat
Pixel format.
#define AV_CEIL_RSHIFT(a, b)