avfilter: add signalstats filter

Signed-off-by: Mark Heath <silicontrip@gmail.com>
Signed-off-by: Dave Rice <dave@dericed.com>
Signed-off-by: Clément Bœsch <u@pkh.me>
pull/76/merge
Clément Bœsch 11 years ago
parent 881ee369e6
commit 76bce46d8f
  1. 1
      Changelog
  2. 166
      doc/filters.texi
  3. 1
      libavfilter/Makefile
  4. 1
      libavfilter/allfilters.c
  5. 2
      libavfilter/version.h
  6. 478
      libavfilter/vf_signalstats.c

@ -28,6 +28,7 @@ version <next>:
- WebVTT encoder - WebVTT encoder
- showcqt multimedia filter - showcqt multimedia filter
- zoompan filter - zoompan filter
- signalstats filter
version 2.2: version 2.2:

@ -7532,6 +7532,172 @@ Swap the second and third planes of the input:
ffmpeg -i INPUT -vf shuffleplanes=0:2:1:3 OUTPUT ffmpeg -i INPUT -vf shuffleplanes=0:2:1:3 OUTPUT
@end example @end example
@section signalstats
Evaluate various visual metrics that assist in determining issues associated
with the digitization of analog video media.
By default the filter will log these metadata values:
@table @option
@item YMIN
Display the minimal Y value contained within the input frame. Expressed in
range of [0-255].
@item YLOW
Display the Y value at the 10% percentile within the input frame. Expressed in
range of [0-255].
@item YAVG
Display the average Y value within the input frame. Expressed in range of
[0-255].
@item YHIGH
Display the Y value at the 90% percentile within the input frame. Expressed in
range of [0-255].
@item YMAX
Display the maximum Y value contained within the input frame. Expressed in
range of [0-255].
@item UMIN
Display the minimal U value contained within the input frame. Expressed in
range of [0-255].
@item ULOW
Display the U value at the 10% percentile within the input frame. Expressed in
range of [0-255].
@item UAVG
Display the average U value within the input frame. Expressed in range of
[0-255].
@item UHIGH
Display the U value at the 90% percentile within the input frame. Expressed in
range of [0-255].
@item UMAX
Display the maximum U value contained within the input frame. Expressed in
range of [0-255].
@item VMIN
Display the minimal V value contained within the input frame. Expressed in
range of [0-255].
@item VLOW
Display the V value at the 10% percentile within the input frame. Expressed in
range of [0-255].
@item VAVG
Display the average V value within the input frame. Expressed in range of
[0-255].
@item VHIGH
Display the V value at the 90% percentile within the input frame. Expressed in
range of [0-255].
@item VMAX
Display the maximum V value contained within the input frame. Expressed in
range of [0-255].
@item SATMIN
Display the minimal saturation value contained within the input frame.
Expressed in range of [0-~181.02].
@item SATLOW
Display the saturation value at the 10% percentile within the input frame.
Expressed in range of [0-~181.02].
@item SATAVG
Display the average saturation value within the input frame. Expressed in range
of [0-~181.02].
@item SATHIGH
Display the saturation value at the 90% percentile within the input frame.
Expressed in range of [0-~181.02].
@item SATMAX
Display the maximum saturation value contained within the input frame.
Expressed in range of [0-~181.02].
@item HUEMED
Display the median value for hue within the input frame. Expressed in range of
[0-360].
@item HUEAVG
Display the average value for hue within the input frame. Expressed in range of
[0-360].
@item YDIF
Display the average of sample value difference between all values of the Y
plane in the current frame and corresponding values of the previous input frame.
Expressed in range of [0-255].
@item UDIF
Display the average of sample value difference between all values of the U
plane in the current frame and corresponding values of the previous input frame.
Expressed in range of [0-255].
@item VDIF
Display the average of sample value difference between all values of the V
plane in the current frame and corresponding values of the previous input frame.
Expressed in range of [0-255].
@end table
The filter accepts the following options:
@table @option
@item stat
@item out
@option{stat} specify an additional form of image analysis.
@option{out} output video with the specified type of pixel highlighted.
Both options accept the following values:
@table @samp
@item tout
Identify @var{temporal outliers} pixels. A @var{temporal outlier} is a pixel
unlike the neighboring pixels of the same field. Examples of temporal outliers
include the results of video dropouts, head clogs, or tape tracking issues.
@item vrep
Identify @var{vertical line repetition}. Vertical line repetition includes
similar rows of pixels within a frame. In born-digital video vertical line
repetition is common, but this pattern is uncommon in video digitized from an
analog source. When it occurs in video that results from the digitization of an
analog source it can indicate concealment from a dropout compensator.
@item brng
Identify pixels that fall outside of legal broadcast range.
@end table
@item color, c
Set the highlight color for the @option{out} option. The default color is
yellow.
@end table
@subsection Examples
@itemize
@item
Output data of various video metrics:
@example
ffprobe -f lavfi movie=example.mov,signalstats="stat=tout+vrep+brng" -show_frames
@end example
@item
Output specific data about the minimum and maximum values of the Y plane per frame:
@example
ffprobe -f lavfi movie=example.mov,signalstats -show_entries frame_tags=lavfi.signalstats.YMAX,lavfi.signalstats.YMIN
@end example
@item
Playback video while highlighting pixels that are outside of broadcast range in red.
@example
ffplay example.mov -vf values="out=brng:color=red"
@end example
@end itemize
@anchor{smartblur} @anchor{smartblur}
@section smartblur @section smartblur

@ -174,6 +174,7 @@ OBJS-$(CONFIG_SETSAR_FILTER) += vf_aspect.o
OBJS-$(CONFIG_SETTB_FILTER) += settb.o OBJS-$(CONFIG_SETTB_FILTER) += settb.o
OBJS-$(CONFIG_SHOWINFO_FILTER) += vf_showinfo.o OBJS-$(CONFIG_SHOWINFO_FILTER) += vf_showinfo.o
OBJS-$(CONFIG_SHUFFLEPLANES_FILTER) += vf_shuffleplanes.o OBJS-$(CONFIG_SHUFFLEPLANES_FILTER) += vf_shuffleplanes.o
OBJS-$(CONFIG_SIGNALSTATS_FILTER) += vf_signalstats.o
OBJS-$(CONFIG_SMARTBLUR_FILTER) += vf_smartblur.o OBJS-$(CONFIG_SMARTBLUR_FILTER) += vf_smartblur.o
OBJS-$(CONFIG_SPLIT_FILTER) += split.o OBJS-$(CONFIG_SPLIT_FILTER) += split.o
OBJS-$(CONFIG_SPP_FILTER) += vf_spp.o OBJS-$(CONFIG_SPP_FILTER) += vf_spp.o

@ -191,6 +191,7 @@ void avfilter_register_all(void)
REGISTER_FILTER(SETTB, settb, vf); REGISTER_FILTER(SETTB, settb, vf);
REGISTER_FILTER(SHOWINFO, showinfo, vf); REGISTER_FILTER(SHOWINFO, showinfo, vf);
REGISTER_FILTER(SHUFFLEPLANES, shuffleplanes, vf); REGISTER_FILTER(SHUFFLEPLANES, shuffleplanes, vf);
REGISTER_FILTER(SIGNALSTATS, signalstats, vf);
REGISTER_FILTER(SMARTBLUR, smartblur, vf); REGISTER_FILTER(SMARTBLUR, smartblur, vf);
REGISTER_FILTER(SPLIT, split, vf); REGISTER_FILTER(SPLIT, split, vf);
REGISTER_FILTER(SPP, spp, vf); REGISTER_FILTER(SPP, spp, vf);

@ -30,7 +30,7 @@
#include "libavutil/version.h" #include "libavutil/version.h"
#define LIBAVFILTER_VERSION_MAJOR 4 #define LIBAVFILTER_VERSION_MAJOR 4
#define LIBAVFILTER_VERSION_MINOR 7 #define LIBAVFILTER_VERSION_MINOR 8
#define LIBAVFILTER_VERSION_MICRO 100 #define LIBAVFILTER_VERSION_MICRO 100
#define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \ #define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \

@ -0,0 +1,478 @@
/*
* Copyright (c) 2010 Mark Heath mjpeg0 @ silicontrip dot org
* Copyright (c) 2014 Clément Bœsch
* Copyright (c) 2014 Dave Rice @dericed
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "internal.h"
enum FilterMode {
FILTER_NONE = -1,
FILTER_TOUT,
FILTER_VREP,
FILTER_BRNG,
FILT_NUMB
};
typedef struct {
const AVClass *class;
int chromah; // height of chroma plane
int chromaw; // width of chroma plane
int hsub; // horizontal subsampling
int vsub; // vertical subsampling
int fs; // pixel count per frame
int cfs; // pixel count per frame of chroma planes
enum FilterMode outfilter;
int filters;
AVFrame *frame_prev;
char *vrep_line;
uint8_t rgba_color[4];
int yuv_color[3];
} SignalstatsContext;
#define OFFSET(x) offsetof(SignalstatsContext, x)
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
static const AVOption signalstats_options[] = {
{"stat", "set statistics filters", OFFSET(filters), AV_OPT_TYPE_FLAGS, {.i64=0}, 0, INT_MAX, FLAGS, "filters"},
{"tout", "analyze pixels for temporal outliers", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_TOUT}, 0, 0, FLAGS, "filters"},
{"vrep", "analyze video lines for vertical line repitition", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_VREP}, 0, 0, FLAGS, "filters"},
{"brng", "analyze for pixels outside of broadcast range", 0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_BRNG}, 0, 0, FLAGS, "filters"},
{"out", "set video filter", OFFSET(outfilter), AV_OPT_TYPE_INT, {.i64=FILTER_NONE}, -1, FILT_NUMB-1, FLAGS, "out"},
{"tout", "highlight pixels that depict temporal outliers", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_TOUT}, 0, 0, FLAGS, "out"},
{"vrep", "highlight video lines that depict vertical line repitition", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_VREP}, 0, 0, FLAGS, "out"},
{"brng", "highlight pixels that are outside of broadcast range", 0, AV_OPT_TYPE_CONST, {.i64=FILTER_BRNG}, 0, 0, FLAGS, "out"},
{"c", "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS},
{"color", "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS},
{NULL}
};
AVFILTER_DEFINE_CLASS(signalstats);
static av_cold int init(AVFilterContext *ctx)
{
uint8_t r, g, b;
SignalstatsContext *s = ctx->priv;
if (s->outfilter != FILTER_NONE)
s->filters |= 1 << s->outfilter;
r = s->rgba_color[0];
g = s->rgba_color[1];
b = s->rgba_color[2];
s->yuv_color[0] = (( 66*r + 129*g + 25*b + (1<<7)) >> 8) + 16;
s->yuv_color[1] = ((-38*r + -74*g + 112*b + (1<<7)) >> 8) + 128;
s->yuv_color[2] = ((112*r + -94*g + -18*b + (1<<7)) >> 8) + 128;
return 0;
}
static av_cold void uninit(AVFilterContext *ctx)
{
SignalstatsContext *s = ctx->priv;
av_frame_free(&s->frame_prev);
av_freep(&s->vrep_line);
}
static int query_formats(AVFilterContext *ctx)
{
// TODO: add more
enum AVPixelFormat pix_fmts[] = {
AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P,
AV_PIX_FMT_NONE
};
ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
return 0;
}
static int config_props(AVFilterLink *outlink)
{
AVFilterContext *ctx = outlink->src;
SignalstatsContext *s = ctx->priv;
AVFilterLink *inlink = outlink->src->inputs[0];
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(outlink->format);
s->hsub = desc->log2_chroma_w;
s->vsub = desc->log2_chroma_h;
outlink->w = inlink->w;
outlink->h = inlink->h;
s->chromaw = FF_CEIL_RSHIFT(inlink->w, s->hsub);
s->chromah = FF_CEIL_RSHIFT(inlink->h, s->vsub);
s->fs = inlink->w * inlink->h;
s->cfs = s->chromaw * s->chromah;
if (s->filters & 1<<FILTER_VREP) {
s->vrep_line = av_malloc(inlink->h * sizeof(*s->vrep_line));
if (!s->vrep_line)
return AVERROR(ENOMEM);
}
return 0;
}
static void burn_frame(SignalstatsContext *s, AVFrame *f, int x, int y)
{
const int chromax = x >> s->hsub;
const int chromay = y >> s->vsub;
f->data[0][y * f->linesize[0] + x] = s->yuv_color[0];
f->data[1][chromay * f->linesize[1] + chromax] = s->yuv_color[1];
f->data[2][chromay * f->linesize[2] + chromax] = s->yuv_color[2];
}
static int filter_brng(SignalstatsContext *s, const AVFrame *in, AVFrame *out, int y, int w, int h)
{
int x, score = 0;
const int yc = y >> s->vsub;
const uint8_t *pluma = &in->data[0][y * in->linesize[0]];
const uint8_t *pchromau = &in->data[1][yc * in->linesize[1]];
const uint8_t *pchromav = &in->data[2][yc * in->linesize[2]];
for (x = 0; x < w; x++) {
const int xc = x >> s->hsub;
const int luma = pluma[x];
const int chromau = pchromau[xc];
const int chromav = pchromav[xc];
const int filt = luma < 16 || luma > 235 ||
chromau < 16 || chromau > 240 ||
chromav < 16 || chromav > 240;
score += filt;
if (out && filt)
burn_frame(s, out, x, y);
}
return score;
}
static int filter_tout_outlier(uint8_t x, uint8_t y, uint8_t z)
{
return ((abs(x - y) + abs (z - y)) / 2) - abs(z - x) > 4; // make 4 configurable?
}
static int filter_tout(SignalstatsContext *s, const AVFrame *in, AVFrame *out, int y, int w, int h)
{
const uint8_t *p = in->data[0];
int lw = in->linesize[0];
int x, score = 0, filt;
if (y - 1 < 0 || y + 1 >= h)
return 0;
// detect two pixels above and below (to eliminate interlace artefacts)
// should check that video format is infact interlaced.
#define FILTER(i, j) \
filter_tout_outlier(p[(y-j) * lw + x + i], \
p[ y * lw + x + i], \
p[(y+j) * lw + x + i])
#define FILTER3(j) (FILTER(-1, j) && FILTER(0, j) && FILTER(1, j))
if (y - 2 >= 0 && y + 2 < h) {
for (x = 1; x < w - 1; x++) {
filt = FILTER3(2) && FILTER3(1);
score += filt;
if (filt && out)
burn_frame(s, out, x, y);
}
} else {
for (x = 1; x < w - 1; x++) {
filt = FILTER3(1);
score += filt;
if (filt && out)
burn_frame(s, out, x, y);
}
}
return score;
}
#define VREP_START 4
static void filter_init_vrep(SignalstatsContext *s, const AVFrame *p, int w, int h)
{
int i, y;
int lw = p->linesize[0];
for (y = VREP_START; y < h; y++) {
int totdiff = 0;
int y2lw = (y - VREP_START) * lw;
int ylw = y * lw;
for (i = 0; i < w; i++)
totdiff += abs(p->data[0][y2lw + i] - p->data[0][ylw + i]);
/* this value should be definable */
s->vrep_line[y] = totdiff < w;
}
}
static int filter_vrep(SignalstatsContext *s, const AVFrame *in, AVFrame *out, int y, int w, int h)
{
int x, score = 0;
if (y < VREP_START)
return 0;
for (x = 0; x < w; x++) {
if (s->vrep_line[y]) {
score++;
if (out)
burn_frame(s, out, x, y);
}
}
return score;
}
static const struct {
const char *name;
void (*init)(SignalstatsContext *s, const AVFrame *p, int w, int h);
int (*process)(SignalstatsContext *s, const AVFrame *in, AVFrame *out, int y, int w, int h);
} filters_def[] = {
{"TOUT", NULL, filter_tout},
{"VREP", filter_init_vrep, filter_vrep},
{"BRNG", NULL, filter_brng},
{NULL}
};
#define DEPTH 256
static int filter_frame(AVFilterLink *link, AVFrame *in)
{
SignalstatsContext *s = link->dst->priv;
AVFilterLink *outlink = link->dst->outputs[0];
AVFrame *out = in;
int i, j;
int w = 0, cw = 0, // in
pw = 0, cpw = 0; // prev
int yuv, yuvu, yuvv;
int fil;
char metabuf[128];
unsigned int histy[DEPTH] = {0},
histu[DEPTH] = {0},
histv[DEPTH] = {0},
histhue[360] = {0},
histsat[DEPTH] = {0}; // limited to 8 bit data.
int miny = -1, minu = -1, minv = -1;
int maxy = -1, maxu = -1, maxv = -1;
int lowy = -1, lowu = -1, lowv = -1;
int highy = -1, highu = -1, highv = -1;
int minsat = -1, maxsat = -1, lowsat = -1, highsat = -1;
int lowp, highp, clowp, chighp;
int accy, accu, accv;
int accsat, acchue = 0;
int medhue, maxhue;
int toty = 0, totu = 0, totv = 0, totsat=0;
int tothue = 0;
int dify = 0, difu = 0, difv = 0;
int filtot[FILT_NUMB] = {0};
AVFrame *prev;
if (!s->frame_prev)
s->frame_prev = av_frame_clone(in);
prev = s->frame_prev;
if (s->outfilter != FILTER_NONE)
out = av_frame_clone(in);
for (fil = 0; fil < FILT_NUMB; fil ++)
if ((s->filters & 1<<fil) && filters_def[fil].init)
filters_def[fil].init(s, in, link->w, link->h);
// Calculate luma histogram and difference with previous frame or field.
for (j = 0; j < link->h; j++) {
for (i = 0; i < link->w; i++) {
yuv = in->data[0][w + i];
histy[yuv]++;
dify += abs(in->data[0][w + i] - prev->data[0][pw + i]);
}
w += in->linesize[0];
pw += prev->linesize[0];
}
// Calculate chroma histogram and difference with previous frame or field.
for (j = 0; j < s->chromah; j++) {
for (i = 0; i < s->chromaw; i++) {
int sat, hue;
yuvu = in->data[1][cw+i];
yuvv = in->data[2][cw+i];
histu[yuvu]++;
difu += abs(in->data[1][cw+i] - prev->data[1][cpw+i]);
histv[yuvv]++;
difv += abs(in->data[2][cw+i] - prev->data[2][cpw+i]);
// int or round?
sat = hypot(yuvu - 128, yuvv - 128);
histsat[sat]++;
hue = floor((180 / M_PI) * atan2f(yuvu-128, yuvv-128) + 180);
histhue[hue]++;
}
cw += in->linesize[1];
cpw += prev->linesize[1];
}
for (j = 0; j < link->h; j++) {
for (fil = 0; fil < FILT_NUMB; fil ++) {
if (s->filters & 1<<fil) {
AVFrame *dbg = out != in && s->outfilter == fil ? out : NULL;
filtot[fil] += filters_def[fil].process(s, in, dbg, j, link->w, link->h);
}
}
}
// find low / high based on histogram percentile
// these only need to be calculated once.
lowp = lrint(s->fs * 10 / 100.);
highp = lrint(s->fs * 90 / 100.);
clowp = lrint(s->cfs * 10 / 100.);
chighp = lrint(s->cfs * 90 / 100.);
accy = accu = accv = accsat = 0;
for (fil = 0; fil < DEPTH; fil++) {
if (miny < 0 && histy[fil]) miny = fil;
if (minu < 0 && histu[fil]) minu = fil;
if (minv < 0 && histv[fil]) minv = fil;
if (minsat < 0 && histsat[fil]) minsat = fil;
if (histy[fil]) maxy = fil;
if (histu[fil]) maxu = fil;
if (histv[fil]) maxv = fil;
if (histsat[fil]) maxsat = fil;
toty += histy[fil] * fil;
totu += histu[fil] * fil;
totv += histv[fil] * fil;
totsat += histsat[fil] * fil;
accy += histy[fil];
accu += histu[fil];
accv += histv[fil];
accsat += histsat[fil];
if (lowy == -1 && accy >= lowp) lowy = fil;
if (lowu == -1 && accu >= clowp) lowu = fil;
if (lowv == -1 && accv >= clowp) lowv = fil;
if (lowsat == -1 && accsat >= clowp) lowsat = fil;
if (highy == -1 && accy >= highp) highy = fil;
if (highu == -1 && accu >= chighp) highu = fil;
if (highv == -1 && accv >= chighp) highv = fil;
if (highsat == -1 && accsat >= chighp) highsat = fil;
}
maxhue = histhue[0];
medhue = -1;
for (fil = 0; fil < 360; fil++) {
tothue += histhue[fil] * fil;
acchue += histhue[fil];
if (medhue == -1 && acchue > s->cfs / 2)
medhue = fil;
if (histhue[fil] > maxhue) {
maxhue = histhue[fil];
}
}
av_frame_free(&s->frame_prev);
s->frame_prev = av_frame_clone(in);
#define SET_META(key, fmt, val) do { \
snprintf(metabuf, sizeof(metabuf), fmt, val); \
av_dict_set(&out->metadata, "lavfi.signalstats." key, metabuf, 0); \
} while (0)
SET_META("YMIN", "%d", miny);
SET_META("YLOW", "%d", lowy);
SET_META("YAVG", "%g", 1.0 * toty / s->fs);
SET_META("YHIGH", "%d", highy);
SET_META("YMAX", "%d", maxy);
SET_META("UMIN", "%d", minu);
SET_META("ULOW", "%d", lowu);
SET_META("UAVG", "%g", 1.0 * totu / s->cfs);
SET_META("UHIGH", "%d", highu);
SET_META("UMAX", "%d", maxu);
SET_META("VMIN", "%d", minv);
SET_META("VLOW", "%d", lowv);
SET_META("VAVG", "%g", 1.0 * totv / s->cfs);
SET_META("VHIGH", "%d", highv);
SET_META("VMAX", "%d", maxv);
SET_META("SATMIN", "%d", minsat);
SET_META("SATLOW", "%d", lowsat);
SET_META("SATAVG", "%g", 1.0 * totsat / s->cfs);
SET_META("SATHIGH", "%d", highsat);
SET_META("SATMAX", "%d", maxsat);
SET_META("HUEMED", "%d", medhue);
SET_META("HUEAVG", "%g", 1.0 * tothue / s->cfs);
SET_META("YDIF", "%g", 1.0 * dify / s->fs);
SET_META("UDIF", "%g", 1.0 * difu / s->cfs);
SET_META("VDIF", "%g", 1.0 * difv / s->cfs);
for (fil = 0; fil < FILT_NUMB; fil ++) {
if (s->filters & 1<<fil) {
char metaname[128];
snprintf(metabuf, sizeof(metabuf), "%g", 1.0 * filtot[fil] / s->fs);
snprintf(metaname, sizeof(metaname), "lavfi.signalstats.%s", filters_def[fil].name);
av_dict_set(&out->metadata, metaname, metabuf, 0);
}
}
if (in != out)
av_frame_free(&in);
return ff_filter_frame(outlink, out);
}
static const AVFilterPad signalstats_inputs[] = {
{
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.filter_frame = filter_frame,
},
{ NULL }
};
static const AVFilterPad signalstats_outputs[] = {
{
.name = "default",
.config_props = config_props,
.type = AVMEDIA_TYPE_VIDEO,
},
{ NULL }
};
AVFilter ff_vf_signalstats = {
.name = "signalstats",
.description = "Generate statistics from video analysis.",
.init = init,
.uninit = uninit,
.query_formats = query_formats,
.priv_size = sizeof(SignalstatsContext),
.inputs = signalstats_inputs,
.outputs = signalstats_outputs,
.priv_class = &signalstats_class,
};
Loading…
Cancel
Save