mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
478 lines
15 KiB
478 lines
15 KiB
/* |
|
* Copyright (c) 2008 Vitor Sessak |
|
* |
|
* This file is part of Libav. |
|
* |
|
* Libav is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* Libav is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with Libav; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
/** |
|
* @file |
|
* memory buffer source filter |
|
*/ |
|
|
|
#include <float.h> |
|
|
|
#include "libavutil/channel_layout.h" |
|
#include "libavutil/common.h" |
|
#include "libavutil/fifo.h" |
|
#include "libavutil/frame.h" |
|
#include "libavutil/imgutils.h" |
|
#include "libavutil/opt.h" |
|
#include "libavutil/samplefmt.h" |
|
#include "audio.h" |
|
#include "avfilter.h" |
|
#include "buffersrc.h" |
|
#include "formats.h" |
|
#include "internal.h" |
|
#include "video.h" |
|
|
|
typedef struct { |
|
const AVClass *class; |
|
AVFifoBuffer *fifo; |
|
AVRational time_base; ///< time_base to set in the output link |
|
|
|
/* video only */ |
|
int h, w; |
|
enum AVPixelFormat pix_fmt; |
|
char *pix_fmt_str; |
|
AVRational pixel_aspect; |
|
|
|
/* audio only */ |
|
int sample_rate; |
|
enum AVSampleFormat sample_fmt; |
|
char *sample_fmt_str; |
|
uint64_t channel_layout; |
|
char *channel_layout_str; |
|
|
|
int eof; |
|
} BufferSourceContext; |
|
|
|
#define CHECK_VIDEO_PARAM_CHANGE(s, c, width, height, format)\ |
|
if (c->w != width || c->h != height || c->pix_fmt != format) {\ |
|
av_log(s, AV_LOG_ERROR, "Changing frame properties on the fly is not supported.\n");\ |
|
return AVERROR(EINVAL);\ |
|
} |
|
|
|
#define CHECK_AUDIO_PARAM_CHANGE(s, c, srate, ch_layout, format)\ |
|
if (c->sample_fmt != format || c->sample_rate != srate ||\ |
|
c->channel_layout != ch_layout) {\ |
|
av_log(s, AV_LOG_ERROR, "Changing frame properties on the fly is not supported.\n");\ |
|
return AVERROR(EINVAL);\ |
|
} |
|
|
|
int av_buffersrc_write_frame(AVFilterContext *ctx, const AVFrame *frame) |
|
{ |
|
AVFrame *copy; |
|
int ret = 0; |
|
|
|
if (!(copy = av_frame_alloc())) |
|
return AVERROR(ENOMEM); |
|
ret = av_frame_ref(copy, frame); |
|
if (ret >= 0) |
|
ret = av_buffersrc_add_frame(ctx, copy); |
|
|
|
av_frame_free(©); |
|
return ret; |
|
} |
|
|
|
int attribute_align_arg av_buffersrc_add_frame(AVFilterContext *ctx, |
|
AVFrame *frame) |
|
{ |
|
BufferSourceContext *s = ctx->priv; |
|
AVFrame *copy; |
|
int ret; |
|
|
|
if (!frame) { |
|
s->eof = 1; |
|
return 0; |
|
} else if (s->eof) |
|
return AVERROR(EINVAL); |
|
|
|
switch (ctx->outputs[0]->type) { |
|
case AVMEDIA_TYPE_VIDEO: |
|
CHECK_VIDEO_PARAM_CHANGE(ctx, s, frame->width, frame->height, |
|
frame->format); |
|
break; |
|
case AVMEDIA_TYPE_AUDIO: |
|
CHECK_AUDIO_PARAM_CHANGE(ctx, s, frame->sample_rate, frame->channel_layout, |
|
frame->format); |
|
break; |
|
default: |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (!av_fifo_space(s->fifo) && |
|
(ret = av_fifo_realloc2(s->fifo, av_fifo_size(s->fifo) + |
|
sizeof(copy))) < 0) |
|
return ret; |
|
|
|
if (!(copy = av_frame_alloc())) |
|
return AVERROR(ENOMEM); |
|
av_frame_move_ref(copy, frame); |
|
|
|
if ((ret = av_fifo_generic_write(s->fifo, ©, sizeof(copy), NULL)) < 0) { |
|
av_frame_move_ref(frame, copy); |
|
av_frame_free(©); |
|
return ret; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
#if FF_API_AVFILTERBUFFER |
|
static void compat_free_buffer(void *opaque, uint8_t *data) |
|
{ |
|
AVFilterBufferRef *buf = opaque; |
|
avfilter_unref_buffer(buf); |
|
} |
|
|
|
static void compat_unref_buffer(void *opaque, uint8_t *data) |
|
{ |
|
AVBufferRef *buf = opaque; |
|
av_buffer_unref(&buf); |
|
} |
|
|
|
int av_buffersrc_buffer(AVFilterContext *ctx, AVFilterBufferRef *buf) |
|
{ |
|
BufferSourceContext *s = ctx->priv; |
|
AVFrame *frame = NULL; |
|
AVBufferRef *dummy_buf = NULL; |
|
int ret = 0, planes, i; |
|
|
|
if (!buf) { |
|
s->eof = 1; |
|
return 0; |
|
} else if (s->eof) |
|
return AVERROR(EINVAL); |
|
|
|
frame = av_frame_alloc(); |
|
if (!frame) |
|
return AVERROR(ENOMEM); |
|
|
|
dummy_buf = av_buffer_create(NULL, 0, compat_free_buffer, buf, 0); |
|
if (!dummy_buf) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
if ((ret = avfilter_copy_buf_props(frame, buf)) < 0) |
|
goto fail; |
|
|
|
#define WRAP_PLANE(ref_out, data, data_size) \ |
|
do { \ |
|
AVBufferRef *dummy_ref = av_buffer_ref(dummy_buf); \ |
|
if (!dummy_ref) { \ |
|
ret = AVERROR(ENOMEM); \ |
|
goto fail; \ |
|
} \ |
|
ref_out = av_buffer_create(data, data_size, compat_unref_buffer, \ |
|
dummy_ref, 0); \ |
|
if (!ref_out) { \ |
|
av_frame_unref(frame); \ |
|
ret = AVERROR(ENOMEM); \ |
|
goto fail; \ |
|
} \ |
|
} while (0) |
|
|
|
if (ctx->outputs[0]->type == AVMEDIA_TYPE_VIDEO) { |
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format); |
|
|
|
planes = av_pix_fmt_count_planes(frame->format); |
|
if (!desc || planes <= 0) { |
|
ret = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
for (i = 0; i < planes; i++) { |
|
int v_shift = (i == 1 || i == 2) ? desc->log2_chroma_h : 0; |
|
int plane_size = (frame->height >> v_shift) * frame->linesize[i]; |
|
|
|
WRAP_PLANE(frame->buf[i], frame->data[i], plane_size); |
|
} |
|
} else { |
|
int planar = av_sample_fmt_is_planar(frame->format); |
|
int channels = av_get_channel_layout_nb_channels(frame->channel_layout); |
|
|
|
planes = planar ? channels : 1; |
|
|
|
if (planes > FF_ARRAY_ELEMS(frame->buf)) { |
|
frame->nb_extended_buf = planes - FF_ARRAY_ELEMS(frame->buf); |
|
frame->extended_buf = av_mallocz(sizeof(*frame->extended_buf) * |
|
frame->nb_extended_buf); |
|
if (!frame->extended_buf) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
} |
|
|
|
for (i = 0; i < FFMIN(planes, FF_ARRAY_ELEMS(frame->buf)); i++) |
|
WRAP_PLANE(frame->buf[i], frame->extended_data[i], frame->linesize[0]); |
|
|
|
for (i = 0; i < planes - FF_ARRAY_ELEMS(frame->buf); i++) |
|
WRAP_PLANE(frame->extended_buf[i], |
|
frame->extended_data[i + FF_ARRAY_ELEMS(frame->buf)], |
|
frame->linesize[0]); |
|
} |
|
|
|
ret = av_buffersrc_add_frame(ctx, frame); |
|
|
|
fail: |
|
av_buffer_unref(&dummy_buf); |
|
av_frame_free(&frame); |
|
|
|
return ret; |
|
} |
|
#endif |
|
|
|
static av_cold int init_video(AVFilterContext *ctx) |
|
{ |
|
BufferSourceContext *c = ctx->priv; |
|
|
|
if (!c->pix_fmt_str || !c->w || !c->h || av_q2d(c->time_base) <= 0) { |
|
av_log(ctx, AV_LOG_ERROR, "Invalid parameters provided.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if ((c->pix_fmt = av_get_pix_fmt(c->pix_fmt_str)) == AV_PIX_FMT_NONE) { |
|
char *tail; |
|
c->pix_fmt = strtol(c->pix_fmt_str, &tail, 10); |
|
if (*tail || c->pix_fmt < 0 || c->pix_fmt >= AV_PIX_FMT_NB) { |
|
av_log(ctx, AV_LOG_ERROR, "Invalid pixel format string '%s'\n", c->pix_fmt_str); |
|
return AVERROR(EINVAL); |
|
} |
|
} |
|
|
|
if (!(c->fifo = av_fifo_alloc(sizeof(AVFrame*)))) |
|
return AVERROR(ENOMEM); |
|
|
|
av_log(ctx, AV_LOG_VERBOSE, "w:%d h:%d pixfmt:%s\n", c->w, c->h, av_get_pix_fmt_name(c->pix_fmt)); |
|
return 0; |
|
} |
|
|
|
#define OFFSET(x) offsetof(BufferSourceContext, x) |
|
#define A AV_OPT_FLAG_AUDIO_PARAM |
|
#define V AV_OPT_FLAG_VIDEO_PARAM |
|
|
|
static const AVOption video_options[] = { |
|
{ "width", NULL, OFFSET(w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
{ "height", NULL, OFFSET(h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
{ "pix_fmt", NULL, OFFSET(pix_fmt_str), AV_OPT_TYPE_STRING, .flags = V }, |
|
#if FF_API_OLD_FILTER_OPTS |
|
/* those 4 are for compatibility with the old option passing system where each filter |
|
* did its own parsing */ |
|
{ "time_base_num", "deprecated, do not use", OFFSET(time_base.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
{ "time_base_den", "deprecated, do not use", OFFSET(time_base.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
{ "sar_num", "deprecated, do not use", OFFSET(pixel_aspect.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
{ "sar_den", "deprecated, do not use", OFFSET(pixel_aspect.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, |
|
#endif |
|
{ "sar", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V }, |
|
{ "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V }, |
|
{ NULL }, |
|
}; |
|
|
|
static const AVClass buffer_class = { |
|
.class_name = "buffer source", |
|
.item_name = av_default_item_name, |
|
.option = video_options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
}; |
|
|
|
static const AVOption audio_options[] = { |
|
{ "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, INT_MAX, A }, |
|
{ "sample_rate", NULL, OFFSET(sample_rate), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A }, |
|
{ "sample_fmt", NULL, OFFSET(sample_fmt_str), AV_OPT_TYPE_STRING, .flags = A }, |
|
{ "channel_layout", NULL, OFFSET(channel_layout_str), AV_OPT_TYPE_STRING, .flags = A }, |
|
{ NULL }, |
|
}; |
|
|
|
static const AVClass abuffer_class = { |
|
.class_name = "abuffer source", |
|
.item_name = av_default_item_name, |
|
.option = audio_options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
}; |
|
|
|
static av_cold int init_audio(AVFilterContext *ctx) |
|
{ |
|
BufferSourceContext *s = ctx->priv; |
|
int ret = 0; |
|
|
|
s->sample_fmt = av_get_sample_fmt(s->sample_fmt_str); |
|
if (s->sample_fmt == AV_SAMPLE_FMT_NONE) { |
|
av_log(ctx, AV_LOG_ERROR, "Invalid sample format %s.\n", |
|
s->sample_fmt_str); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
s->channel_layout = av_get_channel_layout(s->channel_layout_str); |
|
if (!s->channel_layout) { |
|
av_log(ctx, AV_LOG_ERROR, "Invalid channel layout %s.\n", |
|
s->channel_layout_str); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (!(s->fifo = av_fifo_alloc(sizeof(AVFrame*)))) |
|
return AVERROR(ENOMEM); |
|
|
|
if (!s->time_base.num) |
|
s->time_base = (AVRational){1, s->sample_rate}; |
|
|
|
av_log(ctx, AV_LOG_VERBOSE, "tb:%d/%d samplefmt:%s samplerate: %d " |
|
"ch layout:%s\n", s->time_base.num, s->time_base.den, s->sample_fmt_str, |
|
s->sample_rate, s->channel_layout_str); |
|
|
|
return ret; |
|
} |
|
|
|
static av_cold void uninit(AVFilterContext *ctx) |
|
{ |
|
BufferSourceContext *s = ctx->priv; |
|
while (s->fifo && av_fifo_size(s->fifo)) { |
|
AVFrame *frame; |
|
av_fifo_generic_read(s->fifo, &frame, sizeof(frame), NULL); |
|
av_frame_free(&frame); |
|
} |
|
av_fifo_free(s->fifo); |
|
s->fifo = NULL; |
|
} |
|
|
|
static int query_formats(AVFilterContext *ctx) |
|
{ |
|
BufferSourceContext *c = ctx->priv; |
|
AVFilterChannelLayouts *channel_layouts = NULL; |
|
AVFilterFormats *formats = NULL; |
|
AVFilterFormats *samplerates = NULL; |
|
|
|
switch (ctx->outputs[0]->type) { |
|
case AVMEDIA_TYPE_VIDEO: |
|
ff_add_format(&formats, c->pix_fmt); |
|
ff_set_common_formats(ctx, formats); |
|
break; |
|
case AVMEDIA_TYPE_AUDIO: |
|
ff_add_format(&formats, c->sample_fmt); |
|
ff_set_common_formats(ctx, formats); |
|
|
|
ff_add_format(&samplerates, c->sample_rate); |
|
ff_set_common_samplerates(ctx, samplerates); |
|
|
|
ff_add_channel_layout(&channel_layouts, c->channel_layout); |
|
ff_set_common_channel_layouts(ctx, channel_layouts); |
|
break; |
|
default: |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int config_props(AVFilterLink *link) |
|
{ |
|
BufferSourceContext *c = link->src->priv; |
|
|
|
switch (link->type) { |
|
case AVMEDIA_TYPE_VIDEO: |
|
link->w = c->w; |
|
link->h = c->h; |
|
link->sample_aspect_ratio = c->pixel_aspect; |
|
break; |
|
case AVMEDIA_TYPE_AUDIO: |
|
link->channel_layout = c->channel_layout; |
|
link->sample_rate = c->sample_rate; |
|
break; |
|
default: |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
link->time_base = c->time_base; |
|
return 0; |
|
} |
|
|
|
static int request_frame(AVFilterLink *link) |
|
{ |
|
BufferSourceContext *c = link->src->priv; |
|
AVFrame *frame; |
|
int ret = 0; |
|
|
|
if (!av_fifo_size(c->fifo)) { |
|
if (c->eof) |
|
return AVERROR_EOF; |
|
return AVERROR(EAGAIN); |
|
} |
|
av_fifo_generic_read(c->fifo, &frame, sizeof(frame), NULL); |
|
|
|
ff_filter_frame(link, frame); |
|
|
|
return ret; |
|
} |
|
|
|
static int poll_frame(AVFilterLink *link) |
|
{ |
|
BufferSourceContext *c = link->src->priv; |
|
int size = av_fifo_size(c->fifo); |
|
if (!size && c->eof) |
|
return AVERROR_EOF; |
|
return size/sizeof(AVFrame*); |
|
} |
|
|
|
static const AVFilterPad avfilter_vsrc_buffer_outputs[] = { |
|
{ |
|
.name = "default", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.request_frame = request_frame, |
|
.poll_frame = poll_frame, |
|
.config_props = config_props, |
|
}, |
|
{ NULL } |
|
}; |
|
|
|
AVFilter avfilter_vsrc_buffer = { |
|
.name = "buffer", |
|
.description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them accessible to the filterchain."), |
|
.priv_size = sizeof(BufferSourceContext), |
|
.priv_class = &buffer_class, |
|
.query_formats = query_formats, |
|
|
|
.init = init_video, |
|
.uninit = uninit, |
|
|
|
.inputs = NULL, |
|
.outputs = avfilter_vsrc_buffer_outputs, |
|
}; |
|
|
|
static const AVFilterPad avfilter_asrc_abuffer_outputs[] = { |
|
{ |
|
.name = "default", |
|
.type = AVMEDIA_TYPE_AUDIO, |
|
.request_frame = request_frame, |
|
.poll_frame = poll_frame, |
|
.config_props = config_props, |
|
}, |
|
{ NULL } |
|
}; |
|
|
|
AVFilter avfilter_asrc_abuffer = { |
|
.name = "abuffer", |
|
.description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them accessible to the filterchain."), |
|
.priv_size = sizeof(BufferSourceContext), |
|
.priv_class = &abuffer_class, |
|
.query_formats = query_formats, |
|
|
|
.init = init_audio, |
|
.uninit = uninit, |
|
|
|
.inputs = NULL, |
|
.outputs = avfilter_asrc_abuffer_outputs, |
|
};
|
|
|