[FFmpeg-devel] [PATCH] avfilter: add waveform monitor filter
Paul B Mahol
onemda at gmail.com
Wed Aug 19 10:32:01 CEST 2015
Signed-off-by: Paul B Mahol <onemda at gmail.com>
---
doc/filters.texi | 74 +++++++++
libavfilter/Makefile | 1 +
libavfilter/allfilters.c | 1 +
libavfilter/vf_histogram.c | 1 +
libavfilter/vf_waveform.c | 397 +++++++++++++++++++++++++++++++++++++++++++++
5 files changed, 474 insertions(+)
create mode 100644 libavfilter/vf_waveform.c
diff --git a/doc/filters.texi b/doc/filters.texi
index 7b386ef..9d56980 100644
--- a/doc/filters.texi
+++ b/doc/filters.texi
@@ -10885,6 +10885,80 @@ Only deinterlace frames marked as interlaced.
Default value is @samp{all}.
@end table
+ at section waveform
+Video waveform monitor.
+
+The waveform monitor plots color component intensity. By default luminance
+only. Each column of the waveform corresponds to a column of pixels in the
+source video.
+
+It accepts the following options:
+
+ at table @option
+ at item mode, m
+Can be either @code{row}, or @code{column}. Default is @code{column}.
+In row mode, the graph on the left side represents color component value 0 and
+the right side represents value = 255. In column mode, the top side represents
+color component value = 0 and bottom side represents value = 255.
+
+ at item step, s
+Set step. Smaller values are useful to find out how many values of the same
+luminance are distributed across input rows/columns.
+Default value is @code{10}. Allowed range is [1, 255].
+
+ at item mirror, r
+Set mirroring mode. @code{0} means unmirrored, @code{1} means mirrored.
+In mirrored mode, higher values will be represented on the left
+side for @code{row} mode and at the top for @code{column} mode. Default is
+ at code{1} (mirrored).
+
+ at item display, d
+Set display mode.
+It accepts the following values:
+ at table @samp
+ at item overlay
+Presents information identical to that in the @code{parade}, except
+that the graphs representing color components are superimposed directly
+over one another.
+
+This display mode makes it easier to spot relative differences or similarities
+in overlapping areas of the color components that are supposed to be identical,
+such as neutral whites, grays, or blacks.
+
+ at item parade
+Display separate graph for the color components side by side in
+ at code{row} mode or one below the other in @code{column} mode.
+
+Using this display mode makes it easy to spot color casts in the highlights
+and shadows of an image, by comparing the contours of the top and the bottom
+graphs of each waveform. Since whites, grays, and blacks are characterized
+by exactly equal amounts of red, green, and blue, neutral areas of the picture
+should display three waveforms of roughly equal width/height. If not, the
+correction is easy to perform by making level adjustments the three waveforms.
+ at end table
+Default is @code{parade}.
+
+ at item components, c
+Set which color components to display. Default is 1, which means only luminance
+or red color component if input is in RGB colorspace. If is set for example to
+7 it will display all 3 (if) available color components.
+
+ at item envelope, e
+ at table @samp
+ at item none
+No envelope, this is default.
+
+ at item instant
+Instant envelope, minimum and maximum values presented in graph will be easily
+visible even with small @code{step} value.
+
+ at item peak
+Hold minimum and maximum values presented in graph across time. This way you
+can still spot out of range values without constantly looking at waveforms.
+ at end table
+
+ at end table
+
@section xbr
Apply the xBR high-quality magnification filter which is designed for pixel
art. It follows a set of edge-detection rules, see
diff --git a/libavfilter/Makefile b/libavfilter/Makefile
index 75581f2..b15eaf4 100644
--- a/libavfilter/Makefile
+++ b/libavfilter/Makefile
@@ -231,6 +231,7 @@ OBJS-$(CONFIG_VIDSTABDETECT_FILTER) += vidstabutils.o vf_vidstabdetect.
OBJS-$(CONFIG_VIDSTABTRANSFORM_FILTER) += vidstabutils.o vf_vidstabtransform.o
OBJS-$(CONFIG_VIGNETTE_FILTER) += vf_vignette.o
OBJS-$(CONFIG_W3FDIF_FILTER) += vf_w3fdif.o
+OBJS-$(CONFIG_WAVEFORM_FILTER) += vf_waveform.o
OBJS-$(CONFIG_XBR_FILTER) += vf_xbr.o
OBJS-$(CONFIG_YADIF_FILTER) += vf_yadif.o
OBJS-$(CONFIG_ZMQ_FILTER) += f_zmq.o
diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
index ce51382..0756dbb 100644
--- a/libavfilter/allfilters.c
+++ b/libavfilter/allfilters.c
@@ -246,6 +246,7 @@ void avfilter_register_all(void)
REGISTER_FILTER(VIDSTABTRANSFORM, vidstabtransform, vf);
REGISTER_FILTER(VIGNETTE, vignette, vf);
REGISTER_FILTER(W3FDIF, w3fdif, vf);
+ REGISTER_FILTER(WAVEFORM, waveform, vf);
REGISTER_FILTER(XBR, xbr, vf);
REGISTER_FILTER(YADIF, yadif, vf);
REGISTER_FILTER(ZMQ, zmq, vf);
diff --git a/libavfilter/vf_histogram.c b/libavfilter/vf_histogram.c
index 31004b7..7e32b3c 100644
--- a/libavfilter/vf_histogram.c
+++ b/libavfilter/vf_histogram.c
@@ -165,6 +165,7 @@ static int config_output(AVFilterLink *outlink)
outlink->h = (h->level_height + h->scale_height) * FFMAX(h->ncomp * h->display_mode, 1);
break;
case MODE_WAVEFORM:
+ av_log(ctx, AV_LOG_WARNING, "This mode is deprecated, please use waveform filter instead.\n");
if (h->waveform_mode)
outlink->h = 256 * FFMAX(h->ncomp * h->display_mode, 1);
else
diff --git a/libavfilter/vf_waveform.c b/libavfilter/vf_waveform.c
new file mode 100644
index 0000000..fadddeb
--- /dev/null
+++ b/libavfilter/vf_waveform.c
@@ -0,0 +1,397 @@
+/*
+ * Copyright (c) 2012-2015 Paul B Mahol
+ * Copyright (c) 2013 Marton Balint
+ *
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#include "libavutil/avassert.h"
+#include "libavutil/opt.h"
+#include "libavutil/parseutils.h"
+#include "libavutil/pixdesc.h"
+#include "avfilter.h"
+#include "formats.h"
+#include "internal.h"
+#include "video.h"
+
+typedef struct WaveformContext {
+ const AVClass *class;
+ int mode;
+ int ncomp;
+ int pcomp;
+ const uint8_t *bg_color;
+ int step;
+ int mirror;
+ int display;
+ int envelope;
+ int *max[4];
+ int *min[4];
+ const AVPixFmtDescriptor *desc;
+} WaveformContext;
+
+#define OFFSET(x) offsetof(WaveformContext, x)
+#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
+
+static const AVOption waveform_options[] = {
+ { "mode", "set mode", OFFSET(mode), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS, "mode" },
+ { "m", "set mode", OFFSET(mode), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS, "mode" },
+ { "row", NULL, 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, "mode" },
+ { "column", NULL, 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, "mode" },
+ { "step", "set step value", OFFSET(step), AV_OPT_TYPE_INT, {.i64=10}, 1, 255, FLAGS },
+ { "s", "set step value", OFFSET(step), AV_OPT_TYPE_INT, {.i64=10}, 1, 255, FLAGS },
+ { "mirror", "set mirroring", OFFSET(mirror), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS },
+ { "r", "set mirroring", OFFSET(mirror), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS },
+ { "display", "set display mode", OFFSET(display), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS, "display" },
+ { "d", "set display mode", OFFSET(display), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS, "display" },
+ { "overlay", NULL, 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, "display" },
+ { "parade", NULL, 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, "display" },
+ { "components", "set components to display", OFFSET(pcomp), AV_OPT_TYPE_INT, {.i64=1}, 1, 15, FLAGS },
+ { "c", "set components to display", OFFSET(pcomp), AV_OPT_TYPE_INT, {.i64=1}, 1, 15, FLAGS },
+ { "envelope", "set envelope to display", OFFSET(envelope), AV_OPT_TYPE_INT, {.i64=0}, 0, 2, FLAGS, "envelope" },
+ { "e", "set envelope to display", OFFSET(envelope), AV_OPT_TYPE_INT, {.i64=0}, 0, 2, FLAGS, "envelope" },
+ { "none", NULL, 0, AV_OPT_TYPE_CONST, {.i64=0}, 0, 0, FLAGS, "envelope" },
+ { "instant", NULL, 0, AV_OPT_TYPE_CONST, {.i64=1}, 0, 0, FLAGS, "envelope" },
+ { "peak", NULL, 0, AV_OPT_TYPE_CONST, {.i64=2}, 0, 0, FLAGS, "envelope" },
+ { NULL }
+};
+
+AVFILTER_DEFINE_CLASS(waveform);
+
+static const enum AVPixelFormat pix_fmts[] = {
+ AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRAP,
+ AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P,
+ AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P,
+ AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
+ AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUVJ420P,
+ AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P,
+ AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA420P,
+ AV_PIX_FMT_GRAY8,
+ AV_PIX_FMT_NONE
+};
+
+static int query_formats(AVFilterContext *ctx)
+{
+ AVFilterFormats *fmts_list;
+
+ fmts_list = ff_make_format_list(pix_fmts);
+ if (!fmts_list)
+ return AVERROR(ENOMEM);
+ return ff_set_common_formats(ctx, fmts_list);
+}
+
+static const uint8_t black_yuva_color[4] = { 0, 127, 127, 255 };
+static const uint8_t black_gbrp_color[4] = { 0, 0, 0, 255 };
+
+static int config_input(AVFilterLink *inlink)
+{
+ WaveformContext *s = inlink->dst->priv;
+
+ s->desc = av_pix_fmt_desc_get(inlink->format);
+ s->ncomp = s->desc->nb_components;
+
+ switch (inlink->format) {
+ case AV_PIX_FMT_GBRAP:
+ case AV_PIX_FMT_GBRP:
+ s->bg_color = black_gbrp_color;
+ break;
+ default:
+ s->bg_color = black_yuva_color;
+ }
+
+ return 0;
+}
+
+static int config_output(AVFilterLink *outlink)
+{
+ AVFilterContext *ctx = outlink->src;
+ AVFilterLink *inlink = ctx->inputs[0];
+ WaveformContext *s = ctx->priv;
+ int comp = 0, i, p, size, shift;
+
+ for (i = 0; i < s->ncomp; i++) {
+ if ((1 << i) & s->pcomp)
+ comp++;
+ }
+
+ for (p = 0; p < 4; p++) {
+ av_freep(&s->max[p]);
+ av_freep(&s->min[p]);
+ }
+
+ if (s->mode) {
+ outlink->h = 256 * FFMAX(comp * s->display, 1);
+ size = inlink->w * sizeof(int);
+ } else {
+ outlink->w = 256 * FFMAX(comp * s->display, 1);
+ size = inlink->h * sizeof(int);
+ }
+
+ for (p = 0; p < 4; p++) {
+ const int is_chroma = (p == 1 || p == 2);
+ const int shift_w = (is_chroma ? s->desc->log2_chroma_w : 0);
+ const int shift_h = (is_chroma ? s->desc->log2_chroma_h : 0);
+ const int plane = s->desc->comp[p].plane;
+
+ if (!((1 << p) & s->pcomp))
+ continue;
+
+ shift = s->mode ? shift_h : shift_w;
+
+ s->max[plane] = av_malloc(size);
+ s->min[plane] = av_malloc(size);
+
+ if (!s->min[plane] || !s->max[plane])
+ return AVERROR(ENOMEM);
+
+ for (i = 0; i < size / sizeof(int); i++) {
+ s->max[plane][i] = 0;
+ s->min[plane][i] = (s->mode ? outlink->h - 1: outlink->w - 1) >> shift;
+ }
+ }
+
+ outlink->sample_aspect_ratio = (AVRational){1,1};
+
+ return 0;
+}
+
+static void gen_waveform(WaveformContext *s, AVFrame *in, AVFrame *out,
+ int component, int intensity, int offset, int col_mode)
+{
+ const int plane = s->desc->comp[component].plane;
+ const int mirror = s->mirror;
+ const int is_chroma = (component == 1 || component == 2);
+ const int shift_w = (is_chroma ? s->desc->log2_chroma_w : 0);
+ const int shift_h = (is_chroma ? s->desc->log2_chroma_h : 0);
+ const int src_linesize = in->linesize[plane];
+ const int dst_linesize = out->linesize[plane];
+ const int dst_signed_linesize = dst_linesize * (mirror == 1 ? -1 : 1);
+ uint8_t *src_data = in->data[plane];
+ uint8_t *dst_data = out->data[plane] + (col_mode ? (offset >> shift_h) * dst_linesize : offset >> shift_w);
+ uint8_t * const dst_bottom_line = dst_data + dst_linesize * ((256 >> shift_h) - 1);
+ uint8_t * const dst_line = (mirror ? dst_bottom_line : dst_data);
+ const int max = 255 - intensity;
+ const int src_h = FF_CEIL_RSHIFT(in->height, shift_h);
+ const int src_w = FF_CEIL_RSHIFT(in->width, shift_w);
+ uint8_t *dst, *p;
+ int y;
+
+ if (!col_mode && mirror)
+ dst_data += 256 >> shift_w;
+ for (y = 0; y < src_h; y++) {
+ const uint8_t *src_data_end = src_data + src_w;
+ dst = dst_line;
+ for (p = src_data; p < src_data_end; p++) {
+ uint8_t *target;
+ if (col_mode) {
+ target = dst++ + dst_signed_linesize * (*p >> shift_h);
+ } else {
+ if (mirror)
+ target = dst_data - (*p >> shift_w) - 1;
+ else
+ target = dst_data + (*p >> shift_w);
+ }
+ if (*target <= max)
+ *target += intensity;
+ else
+ *target = 255;
+ }
+ src_data += src_linesize;
+ dst_data += dst_linesize;
+ }
+}
+
+static void gen_envelope(WaveformContext *s, AVFrame *out, int component)
+{
+ const int plane = s->desc->comp[component].plane;
+ const int dst_linesize = out->linesize[plane];
+ const uint8_t bg = s->bg_color[plane];
+ const int is_chroma = (component == 1 || component == 2);
+ const int shift_w = (is_chroma ? s->desc->log2_chroma_w : 0);
+ const int shift_h = (is_chroma ? s->desc->log2_chroma_h : 0);
+ const int dst_h = FF_CEIL_RSHIFT(out->height, shift_h);
+ const int dst_w = FF_CEIL_RSHIFT(out->width, shift_w);
+ uint8_t *dst;
+ int x, y;
+
+ if (s->envelope == 0)
+ return;
+
+ if (s->envelope == 1) {
+ if (!s->mode) {
+ for (y = 0; y < dst_h; y++) {
+ dst = out->data[plane] + y * dst_linesize;
+ for (x = 0; x < dst_w; x++) {
+ if (dst[x] != bg) {
+ dst[x] = 255;
+ break;
+ }
+ }
+ for (x = dst_w - 1; x >= 0; x--) {
+ if (dst[x] != bg) {
+ dst[x] = 255;
+ break;
+ }
+ }
+ }
+ } else {
+ for (x = 0; x < dst_w; x++) {
+ for (y = 0; y < dst_h; y++) {
+ dst = out->data[plane] + y * dst_linesize + x;
+ if (dst[0] != bg) {
+ dst[0] = 255;
+ break;
+ }
+ }
+ for (y = dst_h - 1; y >= 0; y--) {
+ dst = out->data[plane] + y * dst_linesize + x;
+ if (dst[0] != bg) {
+ dst[0] = 255;
+ break;
+ }
+ }
+ }
+ }
+ } else {
+ if (!s->mode) {
+ for (y = 0; y < dst_h; y++) {
+ dst = out->data[plane] + y * dst_linesize;
+ for (x = 0; x < dst_w && x < s->min[plane][y]; x++) {
+ if (dst[x] != bg) {
+ s->min[plane][y] = x;
+ break;
+ }
+ }
+ for (x = dst_w - 1; x >= 0 && x >= s->max[plane][y]; x--) {
+ if (dst[x] != bg) {
+ s->max[plane][y] = x;
+ break;
+ }
+ }
+ }
+
+ for (y = 0; y < dst_h; y++) {
+ dst = out->data[plane] + y * dst_linesize + s->min[plane][y];
+ dst[0] = 255;
+ dst = out->data[plane] + y * dst_linesize + s->max[plane][y];
+ dst[0] = 255;
+ }
+ } else {
+ for (x = 0; x < dst_w; x++) {
+ for (y = 0; y < dst_h && y < s->min[plane][x]; y++) {
+ dst = out->data[plane] + y * dst_linesize + x;
+ if (dst[0] != bg) {
+ s->min[plane][x] = y;
+ break;
+ }
+ }
+ for (y = dst_h - 1; y >= 0 && y >= s->max[plane][x]; y--) {
+ dst = out->data[plane] + y * dst_linesize + x;
+ if (dst[0] != bg) {
+ s->max[plane][x] = y;
+ break;
+ }
+ }
+ }
+
+ for (x = 0; x < dst_w; x++) {
+ dst = out->data[plane] + s->min[plane][x] * dst_linesize + x;
+ dst[0] = 255;
+ dst = out->data[plane] + s->max[plane][x] * dst_linesize + x;
+ dst[0] = 255;
+ }
+ }
+ }
+}
+
+static int filter_frame(AVFilterLink *inlink, AVFrame *in)
+{
+ AVFilterContext *ctx = inlink->dst;
+ WaveformContext *s = ctx->priv;
+ AVFilterLink *outlink = ctx->outputs[0];
+ AVFrame *out;
+ int i, k;
+
+ out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
+ if (!out) {
+ av_frame_free(&in);
+ return AVERROR(ENOMEM);
+ }
+
+ out->pts = in->pts;
+
+ for (k = 0; k < s->ncomp; k++) {
+ const int is_chroma = (k == 1 || k == 2);
+ const int dst_h = FF_CEIL_RSHIFT(outlink->h, (is_chroma ? s->desc->log2_chroma_h : 0));
+ const int dst_w = FF_CEIL_RSHIFT(outlink->w, (is_chroma ? s->desc->log2_chroma_w : 0));
+ for (i = 0; i < dst_h ; i++)
+ memset(out->data[s->desc->comp[k].plane] +
+ i * out->linesize[s->desc->comp[k].plane],
+ s->bg_color[k], dst_w);
+ }
+
+ for (k = 0, i = 0; k < s->ncomp; k++) {
+ if ((1 << k) & s->pcomp) {
+ const int offset = i++ * 256 * s->display;
+ gen_waveform(s, in, out, k, s->step, offset, s->mode);
+ gen_envelope(s, out, k);
+ }
+ }
+
+ av_frame_free(&in);
+ return ff_filter_frame(outlink, out);
+}
+
+static av_cold void uninit(AVFilterContext *ctx)
+{
+ WaveformContext *s = ctx->priv;
+ int p;
+
+ for (p = 0; p < 4; p++) {
+ av_freep(&s->max[p]);
+ av_freep(&s->min[p]);
+ }
+}
+
+static const AVFilterPad inputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .filter_frame = filter_frame,
+ .config_props = config_input,
+ },
+ { NULL }
+};
+
+static const AVFilterPad outputs[] = {
+ {
+ .name = "default",
+ .type = AVMEDIA_TYPE_VIDEO,
+ .config_props = config_output,
+ },
+ { NULL }
+};
+
+AVFilter ff_vf_waveform = {
+ .name = "waveform",
+ .description = NULL_IF_CONFIG_SMALL("Video waveform monitor."),
+ .priv_size = sizeof(WaveformContext),
+ .priv_class = &waveform_class,
+ .query_formats = query_formats,
+ .uninit = uninit,
+ .inputs = inputs,
+ .outputs = outputs,
+};
--
1.7.11.2
More information about the ffmpeg-devel
mailing list