mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
465 lines
15 KiB
465 lines
15 KiB
/* |
|
* Copyright (c) 2011 Stefano Sabatini |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
/** |
|
* @file |
|
* buffer sink |
|
*/ |
|
|
|
#include "libavutil/audio_fifo.h" |
|
#include "libavutil/avassert.h" |
|
#include "libavutil/channel_layout.h" |
|
#include "libavutil/common.h" |
|
#include "libavutil/internal.h" |
|
#include "libavutil/mathematics.h" |
|
#include "libavutil/opt.h" |
|
|
|
#define FF_INTERNAL_FIELDS 1 |
|
#include "framequeue.h" |
|
|
|
#include "audio.h" |
|
#include "avfilter.h" |
|
#include "buffersink.h" |
|
#include "internal.h" |
|
|
|
typedef struct BufferSinkContext { |
|
const AVClass *class; |
|
AVFifoBuffer *fifo; ///< FIFO buffer of video frame references |
|
unsigned warning_limit; |
|
|
|
/* only used for video */ |
|
enum AVPixelFormat *pixel_fmts; ///< list of accepted pixel formats, must be terminated with -1 |
|
int pixel_fmts_size; |
|
|
|
/* only used for audio */ |
|
enum AVSampleFormat *sample_fmts; ///< list of accepted sample formats, terminated by AV_SAMPLE_FMT_NONE |
|
int sample_fmts_size; |
|
int64_t *channel_layouts; ///< list of accepted channel layouts, terminated by -1 |
|
int channel_layouts_size; |
|
int *channel_counts; ///< list of accepted channel counts, terminated by -1 |
|
int channel_counts_size; |
|
int all_channel_counts; |
|
int *sample_rates; ///< list of accepted sample rates, terminated by -1 |
|
int sample_rates_size; |
|
|
|
/* only used for compat API */ |
|
AVAudioFifo *audio_fifo; ///< FIFO for audio samples |
|
int64_t next_pts; ///< interpolating audio pts |
|
} BufferSinkContext; |
|
|
|
#define NB_ITEMS(list) (list ## _size / sizeof(*list)) |
|
#define FIFO_INIT_SIZE 8 |
|
#define FIFO_INIT_ELEMENT_SIZE sizeof(void *) |
|
|
|
static av_cold void uninit(AVFilterContext *ctx) |
|
{ |
|
BufferSinkContext *sink = ctx->priv; |
|
AVFrame *frame; |
|
|
|
if (sink->audio_fifo) |
|
av_audio_fifo_free(sink->audio_fifo); |
|
|
|
if (sink->fifo) { |
|
while (av_fifo_size(sink->fifo) >= FIFO_INIT_ELEMENT_SIZE) { |
|
av_fifo_generic_read(sink->fifo, &frame, sizeof(frame), NULL); |
|
av_frame_free(&frame); |
|
} |
|
av_fifo_freep(&sink->fifo); |
|
} |
|
} |
|
|
|
static int add_buffer_ref(AVFilterContext *ctx, AVFrame *ref) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
|
|
if (av_fifo_space(buf->fifo) < FIFO_INIT_ELEMENT_SIZE) { |
|
/* realloc fifo size */ |
|
if (av_fifo_realloc2(buf->fifo, av_fifo_size(buf->fifo) * 2) < 0) { |
|
av_log(ctx, AV_LOG_ERROR, |
|
"Cannot buffer more frames. Consume some available frames " |
|
"before adding new ones.\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
} |
|
|
|
/* cache frame */ |
|
av_fifo_generic_write(buf->fifo, &ref, FIFO_INIT_ELEMENT_SIZE, NULL); |
|
return 0; |
|
} |
|
|
|
static int filter_frame(AVFilterLink *link, AVFrame *frame) |
|
{ |
|
AVFilterContext *ctx = link->dst; |
|
BufferSinkContext *buf = link->dst->priv; |
|
int ret; |
|
|
|
if ((ret = add_buffer_ref(ctx, frame)) < 0) |
|
return ret; |
|
if (buf->warning_limit && |
|
av_fifo_size(buf->fifo) / FIFO_INIT_ELEMENT_SIZE >= buf->warning_limit) { |
|
av_log(ctx, AV_LOG_WARNING, |
|
"%d buffers queued in %s, something may be wrong.\n", |
|
buf->warning_limit, |
|
(char *)av_x_if_null(ctx->name, ctx->filter->name)); |
|
buf->warning_limit *= 10; |
|
} |
|
return 0; |
|
} |
|
|
|
int attribute_align_arg av_buffersink_get_frame(AVFilterContext *ctx, AVFrame *frame) |
|
{ |
|
return av_buffersink_get_frame_flags(ctx, frame, 0); |
|
} |
|
|
|
int attribute_align_arg av_buffersink_get_frame_flags(AVFilterContext *ctx, AVFrame *frame, int flags) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
AVFilterLink *inlink = ctx->inputs[0]; |
|
int peek_in_framequeue = 0, ret; |
|
int64_t frame_count; |
|
AVFrame *cur_frame; |
|
|
|
/* no picref available, fetch it from the filterchain */ |
|
while (!av_fifo_size(buf->fifo)) { |
|
/* if peek_in_framequeue is true later, then ff_request_frame() and |
|
the ff_filter_graph_run_once() loop will take a frame from it and |
|
move it to the internal fifo, ending the global loop */ |
|
av_assert0(!peek_in_framequeue); |
|
if (inlink->status_out) |
|
return inlink->status_out; |
|
peek_in_framequeue = ff_framequeue_queued_frames(&inlink->fifo) && |
|
ff_framequeue_queued_samples(&inlink->fifo) >= inlink->min_samples; |
|
if ((flags & AV_BUFFERSINK_FLAG_NO_REQUEST) && !peek_in_framequeue) |
|
return AVERROR(EAGAIN); |
|
if ((ret = ff_request_frame(inlink)) < 0) |
|
return ret; |
|
frame_count = inlink->frame_count_out; |
|
while (frame_count == inlink->frame_count_out) { |
|
ret = ff_filter_graph_run_once(ctx->graph); |
|
if (ret < 0) |
|
return ret; |
|
} |
|
} |
|
|
|
if (flags & AV_BUFFERSINK_FLAG_PEEK) { |
|
cur_frame = *((AVFrame **)av_fifo_peek2(buf->fifo, 0)); |
|
if ((ret = av_frame_ref(frame, cur_frame)) < 0) |
|
return ret; |
|
} else { |
|
av_fifo_generic_read(buf->fifo, &cur_frame, sizeof(cur_frame), NULL); |
|
av_frame_move_ref(frame, cur_frame); |
|
av_frame_free(&cur_frame); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int read_from_fifo(AVFilterContext *ctx, AVFrame *frame, |
|
int nb_samples) |
|
{ |
|
BufferSinkContext *s = ctx->priv; |
|
AVFilterLink *link = ctx->inputs[0]; |
|
AVFrame *tmp; |
|
|
|
if (!(tmp = ff_get_audio_buffer(link, nb_samples))) |
|
return AVERROR(ENOMEM); |
|
av_audio_fifo_read(s->audio_fifo, (void**)tmp->extended_data, nb_samples); |
|
|
|
tmp->pts = s->next_pts; |
|
if (s->next_pts != AV_NOPTS_VALUE) |
|
s->next_pts += av_rescale_q(nb_samples, (AVRational){1, link->sample_rate}, |
|
link->time_base); |
|
|
|
av_frame_move_ref(frame, tmp); |
|
av_frame_free(&tmp); |
|
|
|
return 0; |
|
} |
|
|
|
int attribute_align_arg av_buffersink_get_samples(AVFilterContext *ctx, |
|
AVFrame *frame, int nb_samples) |
|
{ |
|
BufferSinkContext *s = ctx->priv; |
|
AVFilterLink *link = ctx->inputs[0]; |
|
AVFrame *cur_frame; |
|
int ret = 0; |
|
|
|
if (!s->audio_fifo) { |
|
int nb_channels = link->channels; |
|
if (!(s->audio_fifo = av_audio_fifo_alloc(link->format, nb_channels, nb_samples))) |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
while (ret >= 0) { |
|
if (av_audio_fifo_size(s->audio_fifo) >= nb_samples) |
|
return read_from_fifo(ctx, frame, nb_samples); |
|
|
|
if (!(cur_frame = av_frame_alloc())) |
|
return AVERROR(ENOMEM); |
|
ret = av_buffersink_get_frame_flags(ctx, cur_frame, 0); |
|
if (ret == AVERROR_EOF && av_audio_fifo_size(s->audio_fifo)) { |
|
av_frame_free(&cur_frame); |
|
return read_from_fifo(ctx, frame, av_audio_fifo_size(s->audio_fifo)); |
|
} else if (ret < 0) { |
|
av_frame_free(&cur_frame); |
|
return ret; |
|
} |
|
|
|
if (cur_frame->pts != AV_NOPTS_VALUE) { |
|
s->next_pts = cur_frame->pts - |
|
av_rescale_q(av_audio_fifo_size(s->audio_fifo), |
|
(AVRational){ 1, link->sample_rate }, |
|
link->time_base); |
|
} |
|
|
|
ret = av_audio_fifo_write(s->audio_fifo, (void**)cur_frame->extended_data, |
|
cur_frame->nb_samples); |
|
av_frame_free(&cur_frame); |
|
} |
|
|
|
return ret; |
|
} |
|
|
|
AVBufferSinkParams *av_buffersink_params_alloc(void) |
|
{ |
|
static const int pixel_fmts[] = { AV_PIX_FMT_NONE }; |
|
AVBufferSinkParams *params = av_malloc(sizeof(AVBufferSinkParams)); |
|
if (!params) |
|
return NULL; |
|
|
|
params->pixel_fmts = pixel_fmts; |
|
return params; |
|
} |
|
|
|
AVABufferSinkParams *av_abuffersink_params_alloc(void) |
|
{ |
|
AVABufferSinkParams *params = av_mallocz(sizeof(AVABufferSinkParams)); |
|
|
|
if (!params) |
|
return NULL; |
|
return params; |
|
} |
|
|
|
static av_cold int common_init(AVFilterContext *ctx) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
|
|
buf->fifo = av_fifo_alloc_array(FIFO_INIT_SIZE, FIFO_INIT_ELEMENT_SIZE); |
|
if (!buf->fifo) { |
|
av_log(ctx, AV_LOG_ERROR, "Failed to allocate fifo\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
buf->warning_limit = 100; |
|
buf->next_pts = AV_NOPTS_VALUE; |
|
return 0; |
|
} |
|
|
|
void av_buffersink_set_frame_size(AVFilterContext *ctx, unsigned frame_size) |
|
{ |
|
AVFilterLink *inlink = ctx->inputs[0]; |
|
|
|
inlink->min_samples = inlink->max_samples = |
|
inlink->partial_buf_size = frame_size; |
|
} |
|
|
|
AVRational av_buffersink_get_frame_rate(AVFilterContext *ctx) |
|
{ |
|
av_assert0( !strcmp(ctx->filter->name, "buffersink") |
|
|| !strcmp(ctx->filter->name, "ffbuffersink")); |
|
|
|
return ctx->inputs[0]->frame_rate; |
|
} |
|
|
|
static av_cold int vsink_init(AVFilterContext *ctx, void *opaque) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
AVBufferSinkParams *params = opaque; |
|
int ret; |
|
|
|
if (params) { |
|
if ((ret = av_opt_set_int_list(buf, "pix_fmts", params->pixel_fmts, AV_PIX_FMT_NONE, 0)) < 0) |
|
return ret; |
|
} |
|
|
|
return common_init(ctx); |
|
} |
|
|
|
#define CHECK_LIST_SIZE(field) \ |
|
if (buf->field ## _size % sizeof(*buf->field)) { \ |
|
av_log(ctx, AV_LOG_ERROR, "Invalid size for " #field ": %d, " \ |
|
"should be multiple of %d\n", \ |
|
buf->field ## _size, (int)sizeof(*buf->field)); \ |
|
return AVERROR(EINVAL); \ |
|
} |
|
static int vsink_query_formats(AVFilterContext *ctx) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
AVFilterFormats *formats = NULL; |
|
unsigned i; |
|
int ret; |
|
|
|
CHECK_LIST_SIZE(pixel_fmts) |
|
if (buf->pixel_fmts_size) { |
|
for (i = 0; i < NB_ITEMS(buf->pixel_fmts); i++) |
|
if ((ret = ff_add_format(&formats, buf->pixel_fmts[i])) < 0) |
|
return ret; |
|
if ((ret = ff_set_common_formats(ctx, formats)) < 0) |
|
return ret; |
|
} else { |
|
if ((ret = ff_default_query_formats(ctx)) < 0) |
|
return ret; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static av_cold int asink_init(AVFilterContext *ctx, void *opaque) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
AVABufferSinkParams *params = opaque; |
|
int ret; |
|
|
|
if (params) { |
|
if ((ret = av_opt_set_int_list(buf, "sample_fmts", params->sample_fmts, AV_SAMPLE_FMT_NONE, 0)) < 0 || |
|
(ret = av_opt_set_int_list(buf, "sample_rates", params->sample_rates, -1, 0)) < 0 || |
|
(ret = av_opt_set_int_list(buf, "channel_layouts", params->channel_layouts, -1, 0)) < 0 || |
|
(ret = av_opt_set_int_list(buf, "channel_counts", params->channel_counts, -1, 0)) < 0 || |
|
(ret = av_opt_set_int(buf, "all_channel_counts", params->all_channel_counts, 0)) < 0) |
|
return ret; |
|
} |
|
return common_init(ctx); |
|
} |
|
|
|
static int asink_query_formats(AVFilterContext *ctx) |
|
{ |
|
BufferSinkContext *buf = ctx->priv; |
|
AVFilterFormats *formats = NULL; |
|
AVFilterChannelLayouts *layouts = NULL; |
|
unsigned i; |
|
int ret; |
|
|
|
CHECK_LIST_SIZE(sample_fmts) |
|
CHECK_LIST_SIZE(sample_rates) |
|
CHECK_LIST_SIZE(channel_layouts) |
|
CHECK_LIST_SIZE(channel_counts) |
|
|
|
if (buf->sample_fmts_size) { |
|
for (i = 0; i < NB_ITEMS(buf->sample_fmts); i++) |
|
if ((ret = ff_add_format(&formats, buf->sample_fmts[i])) < 0) |
|
return ret; |
|
if ((ret = ff_set_common_formats(ctx, formats)) < 0) |
|
return ret; |
|
} |
|
|
|
if (buf->channel_layouts_size || buf->channel_counts_size || |
|
buf->all_channel_counts) { |
|
for (i = 0; i < NB_ITEMS(buf->channel_layouts); i++) |
|
if ((ret = ff_add_channel_layout(&layouts, buf->channel_layouts[i])) < 0) |
|
return ret; |
|
for (i = 0; i < NB_ITEMS(buf->channel_counts); i++) |
|
if ((ret = ff_add_channel_layout(&layouts, FF_COUNT2LAYOUT(buf->channel_counts[i]))) < 0) |
|
return ret; |
|
if (buf->all_channel_counts) { |
|
if (layouts) |
|
av_log(ctx, AV_LOG_WARNING, |
|
"Conflicting all_channel_counts and list in options\n"); |
|
else if (!(layouts = ff_all_channel_counts())) |
|
return AVERROR(ENOMEM); |
|
} |
|
if ((ret = ff_set_common_channel_layouts(ctx, layouts)) < 0) |
|
return ret; |
|
} |
|
|
|
if (buf->sample_rates_size) { |
|
formats = NULL; |
|
for (i = 0; i < NB_ITEMS(buf->sample_rates); i++) |
|
if ((ret = ff_add_format(&formats, buf->sample_rates[i])) < 0) |
|
return ret; |
|
if ((ret = ff_set_common_samplerates(ctx, formats)) < 0) |
|
return ret; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
#define OFFSET(x) offsetof(BufferSinkContext, x) |
|
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM |
|
static const AVOption buffersink_options[] = { |
|
{ "pix_fmts", "set the supported pixel formats", OFFSET(pixel_fmts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, |
|
{ NULL }, |
|
}; |
|
#undef FLAGS |
|
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM |
|
static const AVOption abuffersink_options[] = { |
|
{ "sample_fmts", "set the supported sample formats", OFFSET(sample_fmts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, |
|
{ "sample_rates", "set the supported sample rates", OFFSET(sample_rates), AV_OPT_TYPE_BINARY, .flags = FLAGS }, |
|
{ "channel_layouts", "set the supported channel layouts", OFFSET(channel_layouts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, |
|
{ "channel_counts", "set the supported channel counts", OFFSET(channel_counts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, |
|
{ "all_channel_counts", "accept all channel counts", OFFSET(all_channel_counts), AV_OPT_TYPE_BOOL, {.i64 = 0}, 0, 1, FLAGS }, |
|
{ NULL }, |
|
}; |
|
#undef FLAGS |
|
|
|
AVFILTER_DEFINE_CLASS(buffersink); |
|
AVFILTER_DEFINE_CLASS(abuffersink); |
|
|
|
static const AVFilterPad avfilter_vsink_buffer_inputs[] = { |
|
{ |
|
.name = "default", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.filter_frame = filter_frame, |
|
}, |
|
{ NULL } |
|
}; |
|
|
|
AVFilter ff_vsink_buffer = { |
|
.name = "buffersink", |
|
.description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them available to the end of the filter graph."), |
|
.priv_size = sizeof(BufferSinkContext), |
|
.priv_class = &buffersink_class, |
|
.init_opaque = vsink_init, |
|
.uninit = uninit, |
|
|
|
.query_formats = vsink_query_formats, |
|
.inputs = avfilter_vsink_buffer_inputs, |
|
.outputs = NULL, |
|
}; |
|
|
|
static const AVFilterPad avfilter_asink_abuffer_inputs[] = { |
|
{ |
|
.name = "default", |
|
.type = AVMEDIA_TYPE_AUDIO, |
|
.filter_frame = filter_frame, |
|
}, |
|
{ NULL } |
|
}; |
|
|
|
AVFilter ff_asink_abuffer = { |
|
.name = "abuffersink", |
|
.description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them available to the end of the filter graph."), |
|
.priv_class = &abuffersink_class, |
|
.priv_size = sizeof(BufferSinkContext), |
|
.init_opaque = asink_init, |
|
.uninit = uninit, |
|
|
|
.query_formats = asink_query_formats, |
|
.inputs = avfilter_asink_abuffer_inputs, |
|
.outputs = NULL, |
|
};
|
|
|