mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
405 lines
13 KiB
405 lines
13 KiB
/* |
|
* Pulseaudio input |
|
* Copyright (c) 2011 Luca Barbato <lu_zero@gentoo.org> |
|
* Copyright 2004-2006 Lennart Poettering |
|
* Copyright (c) 2014 Michael Niedermayer <michaelni@gmx.at> |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include <pulse/rtclock.h> |
|
#include <pulse/error.h> |
|
|
|
#include "libavutil/internal.h" |
|
#include "libavutil/opt.h" |
|
#include "libavutil/time.h" |
|
|
|
#include "libavformat/avformat.h" |
|
#include "libavformat/internal.h" |
|
#include "libavformat/version.h" |
|
#include "pulse_audio_common.h" |
|
#include "timefilter.h" |
|
|
|
#define DEFAULT_CODEC_ID AV_NE(AV_CODEC_ID_PCM_S16BE, AV_CODEC_ID_PCM_S16LE) |
|
|
|
typedef struct PulseData { |
|
AVClass *class; |
|
char *server; |
|
char *name; |
|
char *stream_name; |
|
int sample_rate; |
|
int channels; |
|
int frame_size; |
|
int fragment_size; |
|
|
|
pa_threaded_mainloop *mainloop; |
|
pa_context *context; |
|
pa_stream *stream; |
|
size_t pa_frame_size; |
|
|
|
TimeFilter *timefilter; |
|
int last_period; |
|
int wallclock; |
|
} PulseData; |
|
|
|
|
|
#define CHECK_SUCCESS_GOTO(rerror, expression, label) \ |
|
do { \ |
|
if (!(expression)) { \ |
|
rerror = AVERROR_EXTERNAL; \ |
|
goto label; \ |
|
} \ |
|
} while (0) |
|
|
|
#define CHECK_DEAD_GOTO(p, rerror, label) \ |
|
do { \ |
|
if (!(p)->context || !PA_CONTEXT_IS_GOOD(pa_context_get_state((p)->context)) || \ |
|
!(p)->stream || !PA_STREAM_IS_GOOD(pa_stream_get_state((p)->stream))) { \ |
|
rerror = AVERROR_EXTERNAL; \ |
|
goto label; \ |
|
} \ |
|
} while (0) |
|
|
|
static void context_state_cb(pa_context *c, void *userdata) { |
|
PulseData *p = userdata; |
|
|
|
switch (pa_context_get_state(c)) { |
|
case PA_CONTEXT_READY: |
|
case PA_CONTEXT_TERMINATED: |
|
case PA_CONTEXT_FAILED: |
|
pa_threaded_mainloop_signal(p->mainloop, 0); |
|
break; |
|
} |
|
} |
|
|
|
static void stream_state_cb(pa_stream *s, void * userdata) { |
|
PulseData *p = userdata; |
|
|
|
switch (pa_stream_get_state(s)) { |
|
case PA_STREAM_READY: |
|
case PA_STREAM_FAILED: |
|
case PA_STREAM_TERMINATED: |
|
pa_threaded_mainloop_signal(p->mainloop, 0); |
|
break; |
|
} |
|
} |
|
|
|
static void stream_request_cb(pa_stream *s, size_t length, void *userdata) { |
|
PulseData *p = userdata; |
|
|
|
pa_threaded_mainloop_signal(p->mainloop, 0); |
|
} |
|
|
|
static void stream_latency_update_cb(pa_stream *s, void *userdata) { |
|
PulseData *p = userdata; |
|
|
|
pa_threaded_mainloop_signal(p->mainloop, 0); |
|
} |
|
|
|
static av_cold int pulse_close(AVFormatContext *s) |
|
{ |
|
PulseData *pd = s->priv_data; |
|
|
|
if (pd->mainloop) |
|
pa_threaded_mainloop_stop(pd->mainloop); |
|
|
|
if (pd->stream) |
|
pa_stream_unref(pd->stream); |
|
pd->stream = NULL; |
|
|
|
if (pd->context) { |
|
pa_context_disconnect(pd->context); |
|
pa_context_unref(pd->context); |
|
} |
|
pd->context = NULL; |
|
|
|
if (pd->mainloop) |
|
pa_threaded_mainloop_free(pd->mainloop); |
|
pd->mainloop = NULL; |
|
|
|
ff_timefilter_destroy(pd->timefilter); |
|
pd->timefilter = NULL; |
|
|
|
return 0; |
|
} |
|
|
|
static av_cold int pulse_read_header(AVFormatContext *s) |
|
{ |
|
PulseData *pd = s->priv_data; |
|
AVStream *st; |
|
char *device = NULL; |
|
int ret; |
|
enum AVCodecID codec_id = |
|
s->audio_codec_id == AV_CODEC_ID_NONE ? DEFAULT_CODEC_ID : s->audio_codec_id; |
|
const pa_sample_spec ss = { ff_codec_id_to_pulse_format(codec_id), |
|
pd->sample_rate, |
|
pd->channels }; |
|
|
|
pa_buffer_attr attr = { -1 }; |
|
pa_channel_map cmap; |
|
const pa_buffer_attr *queried_attr; |
|
|
|
pa_channel_map_init_extend(&cmap, pd->channels, PA_CHANNEL_MAP_WAVEEX); |
|
|
|
st = avformat_new_stream(s, NULL); |
|
|
|
if (!st) { |
|
av_log(s, AV_LOG_ERROR, "Cannot add stream\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
if (pd->fragment_size == -1) { |
|
// 50 ms fragments/latency by default seem good enough |
|
attr.fragsize = pa_frame_size(&ss) * (pd->sample_rate / 20); |
|
} else { |
|
attr.fragsize = pd->fragment_size; |
|
} |
|
|
|
if (s->url[0] != '\0' && strcmp(s->url, "default")) |
|
device = s->url; |
|
|
|
if (!(pd->mainloop = pa_threaded_mainloop_new())) { |
|
pulse_close(s); |
|
return AVERROR_EXTERNAL; |
|
} |
|
|
|
if (!(pd->context = pa_context_new(pa_threaded_mainloop_get_api(pd->mainloop), pd->name))) { |
|
pulse_close(s); |
|
return AVERROR_EXTERNAL; |
|
} |
|
|
|
pa_context_set_state_callback(pd->context, context_state_cb, pd); |
|
|
|
if (pa_context_connect(pd->context, pd->server, 0, NULL) < 0) { |
|
pulse_close(s); |
|
return AVERROR(pa_context_errno(pd->context)); |
|
} |
|
|
|
pa_threaded_mainloop_lock(pd->mainloop); |
|
|
|
if (pa_threaded_mainloop_start(pd->mainloop) < 0) { |
|
ret = -1; |
|
goto unlock_and_fail; |
|
} |
|
|
|
for (;;) { |
|
pa_context_state_t state; |
|
|
|
state = pa_context_get_state(pd->context); |
|
|
|
if (state == PA_CONTEXT_READY) |
|
break; |
|
|
|
if (!PA_CONTEXT_IS_GOOD(state)) { |
|
ret = AVERROR(pa_context_errno(pd->context)); |
|
goto unlock_and_fail; |
|
} |
|
|
|
/* Wait until the context is ready */ |
|
pa_threaded_mainloop_wait(pd->mainloop); |
|
} |
|
|
|
if (!(pd->stream = pa_stream_new(pd->context, pd->stream_name, &ss, &cmap))) { |
|
ret = AVERROR(pa_context_errno(pd->context)); |
|
goto unlock_and_fail; |
|
} |
|
|
|
pa_stream_set_state_callback(pd->stream, stream_state_cb, pd); |
|
pa_stream_set_read_callback(pd->stream, stream_request_cb, pd); |
|
pa_stream_set_write_callback(pd->stream, stream_request_cb, pd); |
|
pa_stream_set_latency_update_callback(pd->stream, stream_latency_update_cb, pd); |
|
|
|
ret = pa_stream_connect_record(pd->stream, device, &attr, |
|
PA_STREAM_INTERPOLATE_TIMING |
|
|PA_STREAM_ADJUST_LATENCY |
|
|PA_STREAM_AUTO_TIMING_UPDATE); |
|
|
|
if (ret < 0) { |
|
ret = AVERROR(pa_context_errno(pd->context)); |
|
goto unlock_and_fail; |
|
} |
|
|
|
for (;;) { |
|
pa_stream_state_t state; |
|
|
|
state = pa_stream_get_state(pd->stream); |
|
|
|
if (state == PA_STREAM_READY) |
|
break; |
|
|
|
if (!PA_STREAM_IS_GOOD(state)) { |
|
ret = AVERROR(pa_context_errno(pd->context)); |
|
goto unlock_and_fail; |
|
} |
|
|
|
/* Wait until the stream is ready */ |
|
pa_threaded_mainloop_wait(pd->mainloop); |
|
} |
|
|
|
/* Query actual fragment size */ |
|
queried_attr = pa_stream_get_buffer_attr(pd->stream); |
|
if (!queried_attr || queried_attr->fragsize > INT_MAX/100) { |
|
ret = AVERROR_EXTERNAL; |
|
goto unlock_and_fail; |
|
} |
|
pd->fragment_size = queried_attr->fragsize; |
|
pd->pa_frame_size = pa_frame_size(&ss); |
|
|
|
pa_threaded_mainloop_unlock(pd->mainloop); |
|
|
|
/* take real parameters */ |
|
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO; |
|
st->codecpar->codec_id = codec_id; |
|
st->codecpar->sample_rate = pd->sample_rate; |
|
st->codecpar->ch_layout.nb_channels = pd->channels; |
|
avpriv_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */ |
|
|
|
pd->timefilter = ff_timefilter_new(1000000.0 / pd->sample_rate, |
|
pd->fragment_size / pd->pa_frame_size, 1.5E-6); |
|
|
|
if (!pd->timefilter) { |
|
pulse_close(s); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
return 0; |
|
|
|
unlock_and_fail: |
|
pa_threaded_mainloop_unlock(pd->mainloop); |
|
|
|
pulse_close(s); |
|
return ret; |
|
} |
|
|
|
static int pulse_read_packet(AVFormatContext *s, AVPacket *pkt) |
|
{ |
|
PulseData *pd = s->priv_data; |
|
int ret; |
|
size_t read_length; |
|
const void *read_data = NULL; |
|
int64_t dts; |
|
pa_usec_t latency; |
|
int negative; |
|
ptrdiff_t pos = 0; |
|
|
|
pa_threaded_mainloop_lock(pd->mainloop); |
|
|
|
CHECK_DEAD_GOTO(pd, ret, unlock_and_fail); |
|
|
|
while (pos < pd->fragment_size) { |
|
int r; |
|
|
|
r = pa_stream_peek(pd->stream, &read_data, &read_length); |
|
CHECK_SUCCESS_GOTO(ret, r == 0, unlock_and_fail); |
|
|
|
if (read_length <= 0) { |
|
pa_threaded_mainloop_wait(pd->mainloop); |
|
CHECK_DEAD_GOTO(pd, ret, unlock_and_fail); |
|
} else if (!read_data) { |
|
/* There's a hole in the stream, skip it. We could generate |
|
* silence, but that wouldn't work for compressed streams. */ |
|
r = pa_stream_drop(pd->stream); |
|
CHECK_SUCCESS_GOTO(ret, r == 0, unlock_and_fail); |
|
} else { |
|
if (!pos) { |
|
if (av_new_packet(pkt, pd->fragment_size) < 0) { |
|
ret = AVERROR(ENOMEM); |
|
goto unlock_and_fail; |
|
} |
|
|
|
dts = av_gettime(); |
|
pa_operation_unref(pa_stream_update_timing_info(pd->stream, NULL, NULL)); |
|
|
|
if (pa_stream_get_latency(pd->stream, &latency, &negative) >= 0) { |
|
if (negative) { |
|
dts += latency; |
|
} else |
|
dts -= latency; |
|
} else { |
|
av_log(s, AV_LOG_WARNING, "pa_stream_get_latency() failed\n"); |
|
} |
|
} |
|
if (pkt->size - pos < read_length) { |
|
if (pos) |
|
break; |
|
pa_stream_drop(pd->stream); |
|
/* Oversized fragment??? */ |
|
ret = AVERROR_EXTERNAL; |
|
goto unlock_and_fail; |
|
} |
|
memcpy(pkt->data + pos, read_data, read_length); |
|
pos += read_length; |
|
pa_stream_drop(pd->stream); |
|
} |
|
} |
|
|
|
pa_threaded_mainloop_unlock(pd->mainloop); |
|
|
|
av_shrink_packet(pkt, pos); |
|
|
|
if (pd->wallclock) |
|
pkt->pts = ff_timefilter_update(pd->timefilter, dts, pd->last_period); |
|
pd->last_period = pkt->size / pd->pa_frame_size; |
|
|
|
return 0; |
|
|
|
unlock_and_fail: |
|
av_packet_unref(pkt); |
|
pa_threaded_mainloop_unlock(pd->mainloop); |
|
return ret; |
|
} |
|
|
|
static int pulse_get_device_list(AVFormatContext *h, AVDeviceInfoList *device_list) |
|
{ |
|
PulseData *s = h->priv_data; |
|
return ff_pulse_audio_get_devices(device_list, s->server, 0); |
|
} |
|
|
|
#define OFFSET(a) offsetof(PulseData, a) |
|
#define D AV_OPT_FLAG_DECODING_PARAM |
|
#define DEPR AV_OPT_FLAG_DEPRECATED |
|
|
|
static const AVOption options[] = { |
|
{ "server", "set PulseAudio server", OFFSET(server), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, D }, |
|
{ "name", "set application name", OFFSET(name), AV_OPT_TYPE_STRING, {.str = LIBAVFORMAT_IDENT}, 0, 0, D }, |
|
{ "stream_name", "set stream description", OFFSET(stream_name), AV_OPT_TYPE_STRING, {.str = "record"}, 0, 0, D }, |
|
{ "sample_rate", "set sample rate in Hz", OFFSET(sample_rate), AV_OPT_TYPE_INT, {.i64 = 48000}, 1, INT_MAX, D }, |
|
{ "channels", "set number of audio channels", OFFSET(channels), AV_OPT_TYPE_INT, {.i64 = 2}, 1, INT_MAX, D }, |
|
{ "frame_size", "set number of bytes per frame", OFFSET(frame_size), AV_OPT_TYPE_INT, {.i64 = 1024}, 1, INT_MAX, D | DEPR }, |
|
{ "fragment_size", "set buffering size, affects latency and cpu usage", OFFSET(fragment_size), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, D }, |
|
{ "wallclock", "set the initial pts using the current time", OFFSET(wallclock), AV_OPT_TYPE_INT, {.i64 = 1}, -1, 1, D }, |
|
{ NULL }, |
|
}; |
|
|
|
static const AVClass pulse_demuxer_class = { |
|
.class_name = "Pulse indev", |
|
.option = options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
.category = AV_CLASS_CATEGORY_DEVICE_AUDIO_INPUT, |
|
}; |
|
|
|
const AVInputFormat ff_pulse_demuxer = { |
|
.name = "pulse", |
|
.long_name = NULL_IF_CONFIG_SMALL("Pulse audio input"), |
|
.priv_data_size = sizeof(PulseData), |
|
.read_header = pulse_read_header, |
|
.read_packet = pulse_read_packet, |
|
.read_close = pulse_close, |
|
.get_device_list = pulse_get_device_list, |
|
.flags = AVFMT_NOFILE, |
|
.priv_class = &pulse_demuxer_class, |
|
};
|
|
|