mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
953 lines
32 KiB
953 lines
32 KiB
/* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include <math.h> |
|
#include <stdint.h> |
|
|
|
#include "ffmpeg.h" |
|
#include "ffmpeg_utils.h" |
|
#include "thread_queue.h" |
|
|
|
#include "libavutil/avassert.h" |
|
#include "libavutil/avstring.h" |
|
#include "libavutil/avutil.h" |
|
#include "libavutil/dict.h" |
|
#include "libavutil/display.h" |
|
#include "libavutil/eval.h" |
|
#include "libavutil/frame.h" |
|
#include "libavutil/intreadwrite.h" |
|
#include "libavutil/log.h" |
|
#include "libavutil/pixdesc.h" |
|
#include "libavutil/rational.h" |
|
#include "libavutil/time.h" |
|
#include "libavutil/timestamp.h" |
|
|
|
#include "libavcodec/avcodec.h" |
|
|
|
#include "libavformat/avformat.h" |
|
|
|
struct Encoder { |
|
// combined size of all the packets received from the encoder |
|
uint64_t data_size; |
|
|
|
// number of packets received from the encoder |
|
uint64_t packets_encoded; |
|
|
|
int opened; |
|
|
|
Scheduler *sch; |
|
unsigned sch_idx; |
|
}; |
|
|
|
// data that is local to the decoder thread and not visible outside of it |
|
typedef struct EncoderThread { |
|
AVFrame *frame; |
|
AVPacket *pkt; |
|
} EncoderThread; |
|
|
|
void enc_free(Encoder **penc) |
|
{ |
|
Encoder *enc = *penc; |
|
|
|
if (!enc) |
|
return; |
|
|
|
av_freep(penc); |
|
} |
|
|
|
int enc_alloc(Encoder **penc, const AVCodec *codec, |
|
Scheduler *sch, unsigned sch_idx) |
|
{ |
|
Encoder *enc; |
|
|
|
*penc = NULL; |
|
|
|
enc = av_mallocz(sizeof(*enc)); |
|
if (!enc) |
|
return AVERROR(ENOMEM); |
|
|
|
enc->sch = sch; |
|
enc->sch_idx = sch_idx; |
|
|
|
*penc = enc; |
|
|
|
return 0; |
|
} |
|
|
|
static int hw_device_setup_for_encode(OutputStream *ost, AVBufferRef *frames_ref) |
|
{ |
|
const AVCodecHWConfig *config; |
|
HWDevice *dev = NULL; |
|
int i; |
|
|
|
if (frames_ref && |
|
((AVHWFramesContext*)frames_ref->data)->format == |
|
ost->enc_ctx->pix_fmt) { |
|
// Matching format, will try to use hw_frames_ctx. |
|
} else { |
|
frames_ref = NULL; |
|
} |
|
|
|
for (i = 0;; i++) { |
|
config = avcodec_get_hw_config(ost->enc_ctx->codec, i); |
|
if (!config) |
|
break; |
|
|
|
if (frames_ref && |
|
config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_FRAMES_CTX && |
|
(config->pix_fmt == AV_PIX_FMT_NONE || |
|
config->pix_fmt == ost->enc_ctx->pix_fmt)) { |
|
av_log(ost->enc_ctx, AV_LOG_VERBOSE, "Using input " |
|
"frames context (format %s) with %s encoder.\n", |
|
av_get_pix_fmt_name(ost->enc_ctx->pix_fmt), |
|
ost->enc_ctx->codec->name); |
|
ost->enc_ctx->hw_frames_ctx = av_buffer_ref(frames_ref); |
|
if (!ost->enc_ctx->hw_frames_ctx) |
|
return AVERROR(ENOMEM); |
|
return 0; |
|
} |
|
|
|
if (!dev && |
|
config->methods & AV_CODEC_HW_CONFIG_METHOD_HW_DEVICE_CTX) |
|
dev = hw_device_get_by_type(config->device_type); |
|
} |
|
|
|
if (dev) { |
|
av_log(ost->enc_ctx, AV_LOG_VERBOSE, "Using device %s " |
|
"(type %s) with %s encoder.\n", dev->name, |
|
av_hwdevice_get_type_name(dev->type), ost->enc_ctx->codec->name); |
|
ost->enc_ctx->hw_device_ctx = av_buffer_ref(dev->device_ref); |
|
if (!ost->enc_ctx->hw_device_ctx) |
|
return AVERROR(ENOMEM); |
|
} else { |
|
// No device required, or no device available. |
|
} |
|
return 0; |
|
} |
|
|
|
static int set_encoder_id(OutputFile *of, OutputStream *ost) |
|
{ |
|
const char *cname = ost->enc_ctx->codec->name; |
|
uint8_t *encoder_string; |
|
int encoder_string_len; |
|
|
|
if (av_dict_get(ost->st->metadata, "encoder", NULL, 0)) |
|
return 0; |
|
|
|
encoder_string_len = sizeof(LIBAVCODEC_IDENT) + strlen(cname) + 2; |
|
encoder_string = av_mallocz(encoder_string_len); |
|
if (!encoder_string) |
|
return AVERROR(ENOMEM); |
|
|
|
if (!of->bitexact && !ost->bitexact) |
|
av_strlcpy(encoder_string, LIBAVCODEC_IDENT " ", encoder_string_len); |
|
else |
|
av_strlcpy(encoder_string, "Lavc ", encoder_string_len); |
|
av_strlcat(encoder_string, cname, encoder_string_len); |
|
av_dict_set(&ost->st->metadata, "encoder", encoder_string, |
|
AV_DICT_DONT_STRDUP_VAL | AV_DICT_DONT_OVERWRITE); |
|
|
|
return 0; |
|
} |
|
|
|
int enc_open(void *opaque, const AVFrame *frame) |
|
{ |
|
OutputStream *ost = opaque; |
|
InputStream *ist = ost->ist; |
|
Encoder *e = ost->enc; |
|
AVCodecContext *enc_ctx = ost->enc_ctx; |
|
AVCodecContext *dec_ctx = NULL; |
|
const AVCodec *enc = enc_ctx->codec; |
|
OutputFile *of = ost->file; |
|
FrameData *fd; |
|
int frame_samples = 0; |
|
int ret; |
|
|
|
if (e->opened) |
|
return 0; |
|
|
|
// frame is always non-NULL for audio and video |
|
av_assert0(frame || (enc->type != AVMEDIA_TYPE_VIDEO && enc->type != AVMEDIA_TYPE_AUDIO)); |
|
|
|
if (frame) { |
|
av_assert0(frame->opaque_ref); |
|
fd = (FrameData*)frame->opaque_ref->data; |
|
} |
|
|
|
ret = set_encoder_id(of, ost); |
|
if (ret < 0) |
|
return ret; |
|
|
|
if (ist) { |
|
dec_ctx = ist->dec_ctx; |
|
} |
|
|
|
// the timebase is chosen by filtering code |
|
if (ost->type == AVMEDIA_TYPE_AUDIO || ost->type == AVMEDIA_TYPE_VIDEO) { |
|
enc_ctx->time_base = frame->time_base; |
|
enc_ctx->framerate = fd->frame_rate_filter; |
|
ost->st->avg_frame_rate = fd->frame_rate_filter; |
|
} |
|
|
|
switch (enc_ctx->codec_type) { |
|
case AVMEDIA_TYPE_AUDIO: |
|
av_assert0(frame->format != AV_SAMPLE_FMT_NONE && |
|
frame->sample_rate > 0 && |
|
frame->ch_layout.nb_channels > 0); |
|
enc_ctx->sample_fmt = frame->format; |
|
enc_ctx->sample_rate = frame->sample_rate; |
|
ret = av_channel_layout_copy(&enc_ctx->ch_layout, &frame->ch_layout); |
|
if (ret < 0) |
|
return ret; |
|
|
|
if (ost->bits_per_raw_sample) |
|
enc_ctx->bits_per_raw_sample = ost->bits_per_raw_sample; |
|
else |
|
enc_ctx->bits_per_raw_sample = FFMIN(fd->bits_per_raw_sample, |
|
av_get_bytes_per_sample(enc_ctx->sample_fmt) << 3); |
|
break; |
|
|
|
case AVMEDIA_TYPE_VIDEO: { |
|
av_assert0(frame->format != AV_PIX_FMT_NONE && |
|
frame->width > 0 && |
|
frame->height > 0); |
|
enc_ctx->width = frame->width; |
|
enc_ctx->height = frame->height; |
|
enc_ctx->sample_aspect_ratio = ost->st->sample_aspect_ratio = |
|
ost->frame_aspect_ratio.num ? // overridden by the -aspect cli option |
|
av_mul_q(ost->frame_aspect_ratio, (AVRational){ enc_ctx->height, enc_ctx->width }) : |
|
frame->sample_aspect_ratio; |
|
|
|
enc_ctx->pix_fmt = frame->format; |
|
|
|
if (ost->bits_per_raw_sample) |
|
enc_ctx->bits_per_raw_sample = ost->bits_per_raw_sample; |
|
else |
|
enc_ctx->bits_per_raw_sample = FFMIN(fd->bits_per_raw_sample, |
|
av_pix_fmt_desc_get(enc_ctx->pix_fmt)->comp[0].depth); |
|
|
|
enc_ctx->color_range = frame->color_range; |
|
enc_ctx->color_primaries = frame->color_primaries; |
|
enc_ctx->color_trc = frame->color_trc; |
|
enc_ctx->colorspace = frame->colorspace; |
|
enc_ctx->chroma_sample_location = frame->chroma_location; |
|
|
|
if (enc_ctx->flags & (AV_CODEC_FLAG_INTERLACED_DCT | AV_CODEC_FLAG_INTERLACED_ME) || |
|
(frame->flags & AV_FRAME_FLAG_INTERLACED) |
|
#if FFMPEG_OPT_TOP |
|
|| ost->top_field_first >= 0 |
|
#endif |
|
) { |
|
int top_field_first = |
|
#if FFMPEG_OPT_TOP |
|
ost->top_field_first >= 0 ? |
|
ost->top_field_first : |
|
#endif |
|
!!(frame->flags & AV_FRAME_FLAG_TOP_FIELD_FIRST); |
|
|
|
if (enc->id == AV_CODEC_ID_MJPEG) |
|
enc_ctx->field_order = top_field_first ? AV_FIELD_TT : AV_FIELD_BB; |
|
else |
|
enc_ctx->field_order = top_field_first ? AV_FIELD_TB : AV_FIELD_BT; |
|
} else |
|
enc_ctx->field_order = AV_FIELD_PROGRESSIVE; |
|
|
|
break; |
|
} |
|
case AVMEDIA_TYPE_SUBTITLE: |
|
if (ost->enc_timebase.num) |
|
av_log(ost, AV_LOG_WARNING, |
|
"-enc_time_base not supported for subtitles, ignoring\n"); |
|
enc_ctx->time_base = AV_TIME_BASE_Q; |
|
|
|
if (!enc_ctx->width) { |
|
enc_ctx->width = ost->ist->par->width; |
|
enc_ctx->height = ost->ist->par->height; |
|
} |
|
if (dec_ctx && dec_ctx->subtitle_header) { |
|
/* ASS code assumes this buffer is null terminated so add extra byte. */ |
|
enc_ctx->subtitle_header = av_mallocz(dec_ctx->subtitle_header_size + 1); |
|
if (!enc_ctx->subtitle_header) |
|
return AVERROR(ENOMEM); |
|
memcpy(enc_ctx->subtitle_header, dec_ctx->subtitle_header, |
|
dec_ctx->subtitle_header_size); |
|
enc_ctx->subtitle_header_size = dec_ctx->subtitle_header_size; |
|
} |
|
|
|
break; |
|
default: |
|
av_assert0(0); |
|
break; |
|
} |
|
|
|
if (ost->bitexact) |
|
enc_ctx->flags |= AV_CODEC_FLAG_BITEXACT; |
|
|
|
if (!av_dict_get(ost->encoder_opts, "threads", NULL, 0)) |
|
av_dict_set(&ost->encoder_opts, "threads", "auto", 0); |
|
|
|
if (enc->capabilities & AV_CODEC_CAP_ENCODER_REORDERED_OPAQUE) { |
|
ret = av_dict_set(&ost->encoder_opts, "flags", "+copy_opaque", AV_DICT_MULTIKEY); |
|
if (ret < 0) |
|
return ret; |
|
} |
|
|
|
av_dict_set(&ost->encoder_opts, "flags", "+frame_duration", AV_DICT_MULTIKEY); |
|
|
|
ret = hw_device_setup_for_encode(ost, frame ? frame->hw_frames_ctx : NULL); |
|
if (ret < 0) { |
|
av_log(ost, AV_LOG_ERROR, |
|
"Encoding hardware device setup failed: %s\n", av_err2str(ret)); |
|
return ret; |
|
} |
|
|
|
if ((ret = avcodec_open2(ost->enc_ctx, enc, &ost->encoder_opts)) < 0) { |
|
if (ret != AVERROR_EXPERIMENTAL) |
|
av_log(ost, AV_LOG_ERROR, "Error while opening encoder - maybe " |
|
"incorrect parameters such as bit_rate, rate, width or height.\n"); |
|
return ret; |
|
} |
|
|
|
e->opened = 1; |
|
|
|
if (ost->enc_ctx->frame_size) |
|
frame_samples = ost->enc_ctx->frame_size; |
|
|
|
ret = check_avoptions(ost->encoder_opts); |
|
if (ret < 0) |
|
return ret; |
|
|
|
if (ost->enc_ctx->bit_rate && ost->enc_ctx->bit_rate < 1000 && |
|
ost->enc_ctx->codec_id != AV_CODEC_ID_CODEC2 /* don't complain about 700 bit/s modes */) |
|
av_log(ost, AV_LOG_WARNING, "The bitrate parameter is set too low." |
|
" It takes bits/s as argument, not kbits/s\n"); |
|
|
|
ret = avcodec_parameters_from_context(ost->par_in, ost->enc_ctx); |
|
if (ret < 0) { |
|
av_log(ost, AV_LOG_FATAL, |
|
"Error initializing the output stream codec context.\n"); |
|
return ret; |
|
} |
|
|
|
/* |
|
* Add global input side data. For now this is naive, and copies it |
|
* from the input stream's global side data. All side data should |
|
* really be funneled over AVFrame and libavfilter, then added back to |
|
* packet side data, and then potentially using the first packet for |
|
* global side data. |
|
*/ |
|
if (ist) { |
|
int i; |
|
for (i = 0; i < ist->st->codecpar->nb_coded_side_data; i++) { |
|
AVPacketSideData *sd_src = &ist->st->codecpar->coded_side_data[i]; |
|
if (sd_src->type != AV_PKT_DATA_CPB_PROPERTIES) { |
|
AVPacketSideData *sd_dst = av_packet_side_data_new(&ost->par_in->coded_side_data, |
|
&ost->par_in->nb_coded_side_data, |
|
sd_src->type, sd_src->size, 0); |
|
if (!sd_dst) |
|
return AVERROR(ENOMEM); |
|
memcpy(sd_dst->data, sd_src->data, sd_src->size); |
|
if (ist->autorotate && sd_src->type == AV_PKT_DATA_DISPLAYMATRIX) |
|
av_display_rotation_set((int32_t *)sd_dst->data, 0); |
|
} |
|
} |
|
} |
|
|
|
// copy timebase while removing common factors |
|
if (ost->st->time_base.num <= 0 || ost->st->time_base.den <= 0) |
|
ost->st->time_base = av_add_q(ost->enc_ctx->time_base, (AVRational){0, 1}); |
|
|
|
ret = of_stream_init(of, ost); |
|
if (ret < 0) |
|
return ret; |
|
|
|
return frame_samples; |
|
} |
|
|
|
static int check_recording_time(OutputStream *ost, int64_t ts, AVRational tb) |
|
{ |
|
OutputFile *of = ost->file; |
|
|
|
if (of->recording_time != INT64_MAX && |
|
av_compare_ts(ts, tb, of->recording_time, AV_TIME_BASE_Q) >= 0) { |
|
return 0; |
|
} |
|
return 1; |
|
} |
|
|
|
static int do_subtitle_out(OutputFile *of, OutputStream *ost, const AVSubtitle *sub, |
|
AVPacket *pkt) |
|
{ |
|
Encoder *e = ost->enc; |
|
int subtitle_out_max_size = 1024 * 1024; |
|
int subtitle_out_size, nb, i, ret; |
|
AVCodecContext *enc; |
|
int64_t pts; |
|
|
|
if (sub->pts == AV_NOPTS_VALUE) { |
|
av_log(ost, AV_LOG_ERROR, "Subtitle packets must have a pts\n"); |
|
return exit_on_error ? AVERROR(EINVAL) : 0; |
|
} |
|
if ((of->start_time != AV_NOPTS_VALUE && sub->pts < of->start_time)) |
|
return 0; |
|
|
|
enc = ost->enc_ctx; |
|
|
|
/* Note: DVB subtitle need one packet to draw them and one other |
|
packet to clear them */ |
|
/* XXX: signal it in the codec context ? */ |
|
if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE) |
|
nb = 2; |
|
else if (enc->codec_id == AV_CODEC_ID_ASS) |
|
nb = FFMAX(sub->num_rects, 1); |
|
else |
|
nb = 1; |
|
|
|
/* shift timestamp to honor -ss and make check_recording_time() work with -t */ |
|
pts = sub->pts; |
|
if (of->start_time != AV_NOPTS_VALUE) |
|
pts -= of->start_time; |
|
for (i = 0; i < nb; i++) { |
|
AVSubtitle local_sub = *sub; |
|
|
|
if (!check_recording_time(ost, pts, AV_TIME_BASE_Q)) |
|
return AVERROR_EOF; |
|
|
|
ret = av_new_packet(pkt, subtitle_out_max_size); |
|
if (ret < 0) |
|
return AVERROR(ENOMEM); |
|
|
|
local_sub.pts = pts; |
|
// start_display_time is required to be 0 |
|
local_sub.pts += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, AV_TIME_BASE_Q); |
|
local_sub.end_display_time -= sub->start_display_time; |
|
local_sub.start_display_time = 0; |
|
|
|
if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE && i == 1) |
|
local_sub.num_rects = 0; |
|
else if (enc->codec_id == AV_CODEC_ID_ASS && sub->num_rects > 0) { |
|
local_sub.num_rects = 1; |
|
local_sub.rects += i; |
|
} |
|
|
|
ost->frames_encoded++; |
|
|
|
subtitle_out_size = avcodec_encode_subtitle(enc, pkt->data, pkt->size, &local_sub); |
|
if (subtitle_out_size < 0) { |
|
av_log(ost, AV_LOG_FATAL, "Subtitle encoding failed\n"); |
|
return subtitle_out_size; |
|
} |
|
|
|
av_shrink_packet(pkt, subtitle_out_size); |
|
pkt->time_base = AV_TIME_BASE_Q; |
|
pkt->pts = sub->pts; |
|
pkt->duration = av_rescale_q(sub->end_display_time, (AVRational){ 1, 1000 }, pkt->time_base); |
|
if (enc->codec_id == AV_CODEC_ID_DVB_SUBTITLE) { |
|
/* XXX: the pts correction is handled here. Maybe handling |
|
it in the codec would be better */ |
|
if (i == 0) |
|
pkt->pts += av_rescale_q(sub->start_display_time, (AVRational){ 1, 1000 }, pkt->time_base); |
|
else |
|
pkt->pts += av_rescale_q(sub->end_display_time, (AVRational){ 1, 1000 }, pkt->time_base); |
|
} |
|
pkt->dts = pkt->pts; |
|
|
|
ret = sch_enc_send(e->sch, e->sch_idx, pkt); |
|
if (ret < 0) { |
|
av_packet_unref(pkt); |
|
return ret; |
|
} |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
void enc_stats_write(OutputStream *ost, EncStats *es, |
|
const AVFrame *frame, const AVPacket *pkt, |
|
uint64_t frame_num) |
|
{ |
|
Encoder *e = ost->enc; |
|
AVIOContext *io = es->io; |
|
AVRational tb = frame ? frame->time_base : pkt->time_base; |
|
int64_t pts = frame ? frame->pts : pkt->pts; |
|
|
|
AVRational tbi = (AVRational){ 0, 1}; |
|
int64_t ptsi = INT64_MAX; |
|
|
|
const FrameData *fd; |
|
|
|
if ((frame && frame->opaque_ref) || (pkt && pkt->opaque_ref)) { |
|
fd = (const FrameData*)(frame ? frame->opaque_ref->data : pkt->opaque_ref->data); |
|
tbi = fd->dec.tb; |
|
ptsi = fd->dec.pts; |
|
} |
|
|
|
pthread_mutex_lock(&es->lock); |
|
|
|
for (size_t i = 0; i < es->nb_components; i++) { |
|
const EncStatsComponent *c = &es->components[i]; |
|
|
|
switch (c->type) { |
|
case ENC_STATS_LITERAL: avio_write (io, c->str, c->str_len); continue; |
|
case ENC_STATS_FILE_IDX: avio_printf(io, "%d", ost->file->index); continue; |
|
case ENC_STATS_STREAM_IDX: avio_printf(io, "%d", ost->index); continue; |
|
case ENC_STATS_TIMEBASE: avio_printf(io, "%d/%d", tb.num, tb.den); continue; |
|
case ENC_STATS_TIMEBASE_IN: avio_printf(io, "%d/%d", tbi.num, tbi.den); continue; |
|
case ENC_STATS_PTS: avio_printf(io, "%"PRId64, pts); continue; |
|
case ENC_STATS_PTS_IN: avio_printf(io, "%"PRId64, ptsi); continue; |
|
case ENC_STATS_PTS_TIME: avio_printf(io, "%g", pts * av_q2d(tb)); continue; |
|
case ENC_STATS_PTS_TIME_IN: avio_printf(io, "%g", ptsi == INT64_MAX ? |
|
INFINITY : ptsi * av_q2d(tbi)); continue; |
|
case ENC_STATS_FRAME_NUM: avio_printf(io, "%"PRIu64, frame_num); continue; |
|
case ENC_STATS_FRAME_NUM_IN: avio_printf(io, "%"PRIu64, fd ? fd->dec.frame_num : -1); continue; |
|
} |
|
|
|
if (frame) { |
|
switch (c->type) { |
|
case ENC_STATS_SAMPLE_NUM: avio_printf(io, "%"PRIu64, ost->samples_encoded); continue; |
|
case ENC_STATS_NB_SAMPLES: avio_printf(io, "%d", frame->nb_samples); continue; |
|
default: av_assert0(0); |
|
} |
|
} else { |
|
switch (c->type) { |
|
case ENC_STATS_DTS: avio_printf(io, "%"PRId64, pkt->dts); continue; |
|
case ENC_STATS_DTS_TIME: avio_printf(io, "%g", pkt->dts * av_q2d(tb)); continue; |
|
case ENC_STATS_PKT_SIZE: avio_printf(io, "%d", pkt->size); continue; |
|
case ENC_STATS_KEYFRAME: avio_write(io, (pkt->flags & AV_PKT_FLAG_KEY) ? |
|
"K" : "N", 1); continue; |
|
case ENC_STATS_BITRATE: { |
|
double duration = FFMAX(pkt->duration, 1) * av_q2d(tb); |
|
avio_printf(io, "%g", 8.0 * pkt->size / duration); |
|
continue; |
|
} |
|
case ENC_STATS_AVG_BITRATE: { |
|
double duration = pkt->dts * av_q2d(tb); |
|
avio_printf(io, "%g", duration > 0 ? 8.0 * e->data_size / duration : -1.); |
|
continue; |
|
} |
|
default: av_assert0(0); |
|
} |
|
} |
|
} |
|
avio_w8(io, '\n'); |
|
avio_flush(io); |
|
|
|
pthread_mutex_unlock(&es->lock); |
|
} |
|
|
|
static inline double psnr(double d) |
|
{ |
|
return -10.0 * log10(d); |
|
} |
|
|
|
static int update_video_stats(OutputStream *ost, const AVPacket *pkt, int write_vstats) |
|
{ |
|
Encoder *e = ost->enc; |
|
const uint8_t *sd = av_packet_get_side_data(pkt, AV_PKT_DATA_QUALITY_STATS, |
|
NULL); |
|
AVCodecContext *enc = ost->enc_ctx; |
|
enum AVPictureType pict_type; |
|
int64_t frame_number; |
|
double ti1, bitrate, avg_bitrate; |
|
double psnr_val = -1; |
|
int quality; |
|
|
|
quality = sd ? AV_RL32(sd) : -1; |
|
pict_type = sd ? sd[4] : AV_PICTURE_TYPE_NONE; |
|
|
|
atomic_store(&ost->quality, quality); |
|
|
|
if ((enc->flags & AV_CODEC_FLAG_PSNR) && sd && sd[5]) { |
|
// FIXME the scaling assumes 8bit |
|
double error = AV_RL64(sd + 8) / (enc->width * enc->height * 255.0 * 255.0); |
|
if (error >= 0 && error <= 1) |
|
psnr_val = psnr(error); |
|
} |
|
|
|
if (!write_vstats) |
|
return 0; |
|
|
|
/* this is executed just the first time update_video_stats is called */ |
|
if (!vstats_file) { |
|
vstats_file = fopen(vstats_filename, "w"); |
|
if (!vstats_file) { |
|
perror("fopen"); |
|
return AVERROR(errno); |
|
} |
|
} |
|
|
|
frame_number = e->packets_encoded; |
|
if (vstats_version <= 1) { |
|
fprintf(vstats_file, "frame= %5"PRId64" q= %2.1f ", frame_number, |
|
quality / (float)FF_QP2LAMBDA); |
|
} else { |
|
fprintf(vstats_file, "out= %2d st= %2d frame= %5"PRId64" q= %2.1f ", |
|
ost->file->index, ost->index, frame_number, |
|
quality / (float)FF_QP2LAMBDA); |
|
} |
|
|
|
if (psnr_val >= 0) |
|
fprintf(vstats_file, "PSNR= %6.2f ", psnr_val); |
|
|
|
fprintf(vstats_file,"f_size= %6d ", pkt->size); |
|
/* compute pts value */ |
|
ti1 = pkt->dts * av_q2d(pkt->time_base); |
|
if (ti1 < 0.01) |
|
ti1 = 0.01; |
|
|
|
bitrate = (pkt->size * 8) / av_q2d(enc->time_base) / 1000.0; |
|
avg_bitrate = (double)(e->data_size * 8) / ti1 / 1000.0; |
|
fprintf(vstats_file, "s_size= %8.0fkB time= %0.3f br= %7.1fkbits/s avg_br= %7.1fkbits/s ", |
|
(double)e->data_size / 1024, ti1, bitrate, avg_bitrate); |
|
fprintf(vstats_file, "type= %c\n", av_get_picture_type_char(pict_type)); |
|
|
|
return 0; |
|
} |
|
|
|
static int encode_frame(OutputFile *of, OutputStream *ost, AVFrame *frame, |
|
AVPacket *pkt) |
|
{ |
|
Encoder *e = ost->enc; |
|
AVCodecContext *enc = ost->enc_ctx; |
|
const char *type_desc = av_get_media_type_string(enc->codec_type); |
|
const char *action = frame ? "encode" : "flush"; |
|
int ret; |
|
|
|
if (frame) { |
|
FrameData *fd = frame_data(frame); |
|
|
|
fd = frame_data(frame); |
|
if (!fd) |
|
return AVERROR(ENOMEM); |
|
|
|
fd->wallclock[LATENCY_PROBE_ENC_PRE] = av_gettime_relative(); |
|
|
|
if (ost->enc_stats_pre.io) |
|
enc_stats_write(ost, &ost->enc_stats_pre, frame, NULL, |
|
ost->frames_encoded); |
|
|
|
ost->frames_encoded++; |
|
ost->samples_encoded += frame->nb_samples; |
|
|
|
if (debug_ts) { |
|
av_log(ost, AV_LOG_INFO, "encoder <- type:%s " |
|
"frame_pts:%s frame_pts_time:%s time_base:%d/%d\n", |
|
type_desc, |
|
av_ts2str(frame->pts), av_ts2timestr(frame->pts, &enc->time_base), |
|
enc->time_base.num, enc->time_base.den); |
|
} |
|
|
|
if (frame->sample_aspect_ratio.num && !ost->frame_aspect_ratio.num) |
|
enc->sample_aspect_ratio = frame->sample_aspect_ratio; |
|
} |
|
|
|
update_benchmark(NULL); |
|
|
|
ret = avcodec_send_frame(enc, frame); |
|
if (ret < 0 && !(ret == AVERROR_EOF && !frame)) { |
|
av_log(ost, AV_LOG_ERROR, "Error submitting %s frame to the encoder\n", |
|
type_desc); |
|
return ret; |
|
} |
|
|
|
while (1) { |
|
FrameData *fd; |
|
|
|
av_packet_unref(pkt); |
|
|
|
ret = avcodec_receive_packet(enc, pkt); |
|
update_benchmark("%s_%s %d.%d", action, type_desc, |
|
of->index, ost->index); |
|
|
|
pkt->time_base = enc->time_base; |
|
|
|
/* if two pass, output log on success and EOF */ |
|
if ((ret >= 0 || ret == AVERROR_EOF) && ost->logfile && enc->stats_out) |
|
fprintf(ost->logfile, "%s", enc->stats_out); |
|
|
|
if (ret == AVERROR(EAGAIN)) { |
|
av_assert0(frame); // should never happen during flushing |
|
return 0; |
|
} else if (ret < 0) { |
|
if (ret != AVERROR_EOF) |
|
av_log(ost, AV_LOG_ERROR, "%s encoding failed\n", type_desc); |
|
return ret; |
|
} |
|
|
|
fd = packet_data(pkt); |
|
if (!fd) |
|
return AVERROR(ENOMEM); |
|
fd->wallclock[LATENCY_PROBE_ENC_POST] = av_gettime_relative(); |
|
|
|
if (enc->codec_type == AVMEDIA_TYPE_VIDEO) { |
|
ret = update_video_stats(ost, pkt, !!vstats_filename); |
|
if (ret < 0) |
|
return ret; |
|
} |
|
|
|
if (ost->enc_stats_post.io) |
|
enc_stats_write(ost, &ost->enc_stats_post, NULL, pkt, |
|
e->packets_encoded); |
|
|
|
if (debug_ts) { |
|
av_log(ost, AV_LOG_INFO, "encoder -> type:%s " |
|
"pkt_pts:%s pkt_pts_time:%s pkt_dts:%s pkt_dts_time:%s " |
|
"duration:%s duration_time:%s\n", |
|
type_desc, |
|
av_ts2str(pkt->pts), av_ts2timestr(pkt->pts, &enc->time_base), |
|
av_ts2str(pkt->dts), av_ts2timestr(pkt->dts, &enc->time_base), |
|
av_ts2str(pkt->duration), av_ts2timestr(pkt->duration, &enc->time_base)); |
|
} |
|
|
|
e->data_size += pkt->size; |
|
|
|
e->packets_encoded++; |
|
|
|
ret = sch_enc_send(e->sch, e->sch_idx, pkt); |
|
if (ret < 0) { |
|
av_packet_unref(pkt); |
|
return ret; |
|
} |
|
} |
|
|
|
av_assert0(0); |
|
} |
|
|
|
static int do_audio_out(OutputFile *of, OutputStream *ost, |
|
AVFrame *frame, AVPacket *pkt) |
|
{ |
|
AVCodecContext *enc = ost->enc_ctx; |
|
|
|
if (!(enc->codec->capabilities & AV_CODEC_CAP_PARAM_CHANGE) && |
|
enc->ch_layout.nb_channels != frame->ch_layout.nb_channels) { |
|
av_log(ost, AV_LOG_ERROR, |
|
"Audio channel count changed and encoder does not support parameter changes\n"); |
|
return 0; |
|
} |
|
|
|
if (!check_recording_time(ost, frame->pts, frame->time_base)) |
|
return AVERROR_EOF; |
|
|
|
return encode_frame(of, ost, frame, pkt); |
|
} |
|
|
|
static enum AVPictureType forced_kf_apply(void *logctx, KeyframeForceCtx *kf, |
|
AVRational tb, const AVFrame *in_picture) |
|
{ |
|
double pts_time; |
|
|
|
if (kf->ref_pts == AV_NOPTS_VALUE) |
|
kf->ref_pts = in_picture->pts; |
|
|
|
pts_time = (in_picture->pts - kf->ref_pts) * av_q2d(tb); |
|
if (kf->index < kf->nb_pts && |
|
av_compare_ts(in_picture->pts, tb, kf->pts[kf->index], AV_TIME_BASE_Q) >= 0) { |
|
kf->index++; |
|
goto force_keyframe; |
|
} else if (kf->pexpr) { |
|
double res; |
|
kf->expr_const_values[FKF_T] = pts_time; |
|
res = av_expr_eval(kf->pexpr, |
|
kf->expr_const_values, NULL); |
|
av_log(logctx, AV_LOG_TRACE, |
|
"force_key_frame: n:%f n_forced:%f prev_forced_n:%f t:%f prev_forced_t:%f -> res:%f\n", |
|
kf->expr_const_values[FKF_N], |
|
kf->expr_const_values[FKF_N_FORCED], |
|
kf->expr_const_values[FKF_PREV_FORCED_N], |
|
kf->expr_const_values[FKF_T], |
|
kf->expr_const_values[FKF_PREV_FORCED_T], |
|
res); |
|
|
|
kf->expr_const_values[FKF_N] += 1; |
|
|
|
if (res) { |
|
kf->expr_const_values[FKF_PREV_FORCED_N] = kf->expr_const_values[FKF_N] - 1; |
|
kf->expr_const_values[FKF_PREV_FORCED_T] = kf->expr_const_values[FKF_T]; |
|
kf->expr_const_values[FKF_N_FORCED] += 1; |
|
goto force_keyframe; |
|
} |
|
} else if (kf->type == KF_FORCE_SOURCE && (in_picture->flags & AV_FRAME_FLAG_KEY)) { |
|
goto force_keyframe; |
|
} |
|
|
|
return AV_PICTURE_TYPE_NONE; |
|
|
|
force_keyframe: |
|
av_log(logctx, AV_LOG_DEBUG, "Forced keyframe at time %f\n", pts_time); |
|
return AV_PICTURE_TYPE_I; |
|
} |
|
|
|
/* May modify/reset frame */ |
|
static int do_video_out(OutputFile *of, OutputStream *ost, |
|
AVFrame *in_picture, AVPacket *pkt) |
|
{ |
|
AVCodecContext *enc = ost->enc_ctx; |
|
|
|
if (!check_recording_time(ost, in_picture->pts, ost->enc_ctx->time_base)) |
|
return AVERROR_EOF; |
|
|
|
in_picture->quality = enc->global_quality; |
|
in_picture->pict_type = forced_kf_apply(ost, &ost->kf, enc->time_base, in_picture); |
|
|
|
#if FFMPEG_OPT_TOP |
|
if (ost->top_field_first >= 0) { |
|
in_picture->flags &= ~AV_FRAME_FLAG_TOP_FIELD_FIRST; |
|
in_picture->flags |= AV_FRAME_FLAG_TOP_FIELD_FIRST * (!!ost->top_field_first); |
|
} |
|
#endif |
|
|
|
return encode_frame(of, ost, in_picture, pkt); |
|
} |
|
|
|
static int frame_encode(OutputStream *ost, AVFrame *frame, AVPacket *pkt) |
|
{ |
|
OutputFile *of = ost->file; |
|
enum AVMediaType type = ost->type; |
|
|
|
if (type == AVMEDIA_TYPE_SUBTITLE) { |
|
const AVSubtitle *subtitle = frame && frame->buf[0] ? |
|
(AVSubtitle*)frame->buf[0]->data : NULL; |
|
|
|
// no flushing for subtitles |
|
return subtitle && subtitle->num_rects ? |
|
do_subtitle_out(of, ost, subtitle, pkt) : 0; |
|
} |
|
|
|
if (frame) { |
|
return (type == AVMEDIA_TYPE_VIDEO) ? do_video_out(of, ost, frame, pkt) : |
|
do_audio_out(of, ost, frame, pkt); |
|
} |
|
|
|
return encode_frame(of, ost, NULL, pkt); |
|
} |
|
|
|
static void enc_thread_set_name(const OutputStream *ost) |
|
{ |
|
char name[16]; |
|
snprintf(name, sizeof(name), "enc%d:%d:%s", ost->file->index, ost->index, |
|
ost->enc_ctx->codec->name); |
|
ff_thread_setname(name); |
|
} |
|
|
|
static void enc_thread_uninit(EncoderThread *et) |
|
{ |
|
av_packet_free(&et->pkt); |
|
av_frame_free(&et->frame); |
|
|
|
memset(et, 0, sizeof(*et)); |
|
} |
|
|
|
static int enc_thread_init(EncoderThread *et) |
|
{ |
|
memset(et, 0, sizeof(*et)); |
|
|
|
et->frame = av_frame_alloc(); |
|
if (!et->frame) |
|
goto fail; |
|
|
|
et->pkt = av_packet_alloc(); |
|
if (!et->pkt) |
|
goto fail; |
|
|
|
return 0; |
|
|
|
fail: |
|
enc_thread_uninit(et); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
void *encoder_thread(void *arg) |
|
{ |
|
OutputStream *ost = arg; |
|
Encoder *e = ost->enc; |
|
EncoderThread et; |
|
int ret = 0, input_status = 0; |
|
int name_set = 0; |
|
|
|
ret = enc_thread_init(&et); |
|
if (ret < 0) |
|
goto finish; |
|
|
|
/* Open the subtitle encoders immediately. AVFrame-based encoders |
|
* are opened through a callback from the scheduler once they get |
|
* their first frame |
|
* |
|
* N.B.: because the callback is called from a different thread, |
|
* enc_ctx MUST NOT be accessed before sch_enc_receive() returns |
|
* for the first time for audio/video. */ |
|
if (ost->type != AVMEDIA_TYPE_VIDEO && ost->type != AVMEDIA_TYPE_AUDIO) { |
|
ret = enc_open(ost, NULL); |
|
if (ret < 0) |
|
goto finish; |
|
} |
|
|
|
while (!input_status) { |
|
input_status = sch_enc_receive(e->sch, e->sch_idx, et.frame); |
|
if (input_status < 0) { |
|
if (input_status == AVERROR_EOF) { |
|
av_log(ost, AV_LOG_VERBOSE, "Encoder thread received EOF\n"); |
|
if (e->opened) |
|
break; |
|
|
|
av_log(ost, AV_LOG_ERROR, "Could not open encoder before EOF\n"); |
|
ret = AVERROR(EINVAL); |
|
} else { |
|
av_log(ost, AV_LOG_ERROR, "Error receiving a frame for encoding: %s\n", |
|
av_err2str(ret)); |
|
ret = input_status; |
|
} |
|
goto finish; |
|
} |
|
|
|
if (!name_set) { |
|
enc_thread_set_name(ost); |
|
name_set = 1; |
|
} |
|
|
|
ret = frame_encode(ost, et.frame, et.pkt); |
|
|
|
av_packet_unref(et.pkt); |
|
av_frame_unref(et.frame); |
|
|
|
if (ret < 0) { |
|
if (ret == AVERROR_EOF) |
|
av_log(ost, AV_LOG_VERBOSE, "Encoder returned EOF, finishing\n"); |
|
else |
|
av_log(ost, AV_LOG_ERROR, "Error encoding a frame: %s\n", |
|
av_err2str(ret)); |
|
break; |
|
} |
|
} |
|
|
|
// flush the encoder |
|
if (ret == 0 || ret == AVERROR_EOF) { |
|
ret = frame_encode(ost, NULL, et.pkt); |
|
if (ret < 0 && ret != AVERROR_EOF) |
|
av_log(ost, AV_LOG_ERROR, "Error flushing encoder: %s\n", |
|
av_err2str(ret)); |
|
} |
|
|
|
// EOF is normal thread termination |
|
if (ret == AVERROR_EOF) |
|
ret = 0; |
|
|
|
finish: |
|
enc_thread_uninit(&et); |
|
|
|
return (void*)(intptr_t)ret; |
|
}
|
|
|