mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1584 lines
61 KiB
1584 lines
61 KiB
/* |
|
* Copyright (c) 2010, Google, Inc. |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
/** |
|
* @file |
|
* VP8/9 encoder support via libvpx |
|
*/ |
|
|
|
#define VPX_DISABLE_CTRL_TYPECHECKS 1 |
|
#define VPX_CODEC_DISABLE_COMPAT 1 |
|
#include <vpx/vpx_encoder.h> |
|
#include <vpx/vp8cx.h> |
|
|
|
#include "avcodec.h" |
|
#include "internal.h" |
|
#include "libavutil/avassert.h" |
|
#include "libvpx.h" |
|
#include "profiles.h" |
|
#include "libavutil/avstring.h" |
|
#include "libavutil/base64.h" |
|
#include "libavutil/common.h" |
|
#include "libavutil/internal.h" |
|
#include "libavutil/intreadwrite.h" |
|
#include "libavutil/mathematics.h" |
|
#include "libavutil/opt.h" |
|
|
|
/** |
|
* Portion of struct vpx_codec_cx_pkt from vpx_encoder.h. |
|
* One encoded frame returned from the library. |
|
*/ |
|
struct FrameListData { |
|
void *buf; /**< compressed data buffer */ |
|
size_t sz; /**< length of compressed data */ |
|
void *buf_alpha; |
|
size_t sz_alpha; |
|
int64_t pts; /**< time stamp to show frame |
|
(in timebase units) */ |
|
unsigned long duration; /**< duration to show frame |
|
(in timebase units) */ |
|
uint32_t flags; /**< flags for this frame */ |
|
uint64_t sse[4]; |
|
int have_sse; /**< true if we have pending sse[] */ |
|
uint64_t frame_number; |
|
struct FrameListData *next; |
|
}; |
|
|
|
typedef struct VPxEncoderContext { |
|
AVClass *class; |
|
struct vpx_codec_ctx encoder; |
|
struct vpx_image rawimg; |
|
struct vpx_codec_ctx encoder_alpha; |
|
struct vpx_image rawimg_alpha; |
|
uint8_t is_alpha; |
|
struct vpx_fixed_buf twopass_stats; |
|
int deadline; //i.e., RT/GOOD/BEST |
|
uint64_t sse[4]; |
|
int have_sse; /**< true if we have pending sse[] */ |
|
uint64_t frame_number; |
|
struct FrameListData *coded_frame_list; |
|
|
|
int cpu_used; |
|
int sharpness; |
|
/** |
|
* VP8 specific flags, see VP8F_* below. |
|
*/ |
|
int flags; |
|
#define VP8F_ERROR_RESILIENT 0x00000001 ///< Enable measures appropriate for streaming over lossy links |
|
#define VP8F_AUTO_ALT_REF 0x00000002 ///< Enable automatic alternate reference frame generation |
|
|
|
int auto_alt_ref; |
|
|
|
int arnr_max_frames; |
|
int arnr_strength; |
|
int arnr_type; |
|
|
|
int tune; |
|
|
|
int lag_in_frames; |
|
int error_resilient; |
|
int crf; |
|
int static_thresh; |
|
int max_intra_rate; |
|
int rc_undershoot_pct; |
|
int rc_overshoot_pct; |
|
|
|
AVDictionary *vp8_ts_parameters; |
|
|
|
// VP9-only |
|
int lossless; |
|
int tile_columns; |
|
int tile_rows; |
|
int frame_parallel; |
|
int aq_mode; |
|
int drop_threshold; |
|
int noise_sensitivity; |
|
int vpx_cs; |
|
float level; |
|
int row_mt; |
|
int tune_content; |
|
int corpus_complexity; |
|
int tpl_model; |
|
/** |
|
* If the driver does not support ROI then warn the first time we |
|
* encounter a frame with ROI side data. |
|
*/ |
|
int roi_warned; |
|
} VPxContext; |
|
|
|
/** String mappings for enum vp8e_enc_control_id */ |
|
static const char *const ctlidstr[] = { |
|
[VP8E_SET_CPUUSED] = "VP8E_SET_CPUUSED", |
|
[VP8E_SET_ENABLEAUTOALTREF] = "VP8E_SET_ENABLEAUTOALTREF", |
|
[VP8E_SET_NOISE_SENSITIVITY] = "VP8E_SET_NOISE_SENSITIVITY", |
|
[VP8E_SET_STATIC_THRESHOLD] = "VP8E_SET_STATIC_THRESHOLD", |
|
[VP8E_SET_TOKEN_PARTITIONS] = "VP8E_SET_TOKEN_PARTITIONS", |
|
[VP8E_SET_ARNR_MAXFRAMES] = "VP8E_SET_ARNR_MAXFRAMES", |
|
[VP8E_SET_ARNR_STRENGTH] = "VP8E_SET_ARNR_STRENGTH", |
|
[VP8E_SET_ARNR_TYPE] = "VP8E_SET_ARNR_TYPE", |
|
[VP8E_SET_TUNING] = "VP8E_SET_TUNING", |
|
[VP8E_SET_CQ_LEVEL] = "VP8E_SET_CQ_LEVEL", |
|
[VP8E_SET_MAX_INTRA_BITRATE_PCT] = "VP8E_SET_MAX_INTRA_BITRATE_PCT", |
|
[VP8E_SET_SHARPNESS] = "VP8E_SET_SHARPNESS", |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
[VP9E_SET_LOSSLESS] = "VP9E_SET_LOSSLESS", |
|
[VP9E_SET_TILE_COLUMNS] = "VP9E_SET_TILE_COLUMNS", |
|
[VP9E_SET_TILE_ROWS] = "VP9E_SET_TILE_ROWS", |
|
[VP9E_SET_FRAME_PARALLEL_DECODING] = "VP9E_SET_FRAME_PARALLEL_DECODING", |
|
[VP9E_SET_AQ_MODE] = "VP9E_SET_AQ_MODE", |
|
[VP9E_SET_COLOR_SPACE] = "VP9E_SET_COLOR_SPACE", |
|
#if VPX_ENCODER_ABI_VERSION >= 11 |
|
[VP9E_SET_COLOR_RANGE] = "VP9E_SET_COLOR_RANGE", |
|
#endif |
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
[VP9E_SET_TARGET_LEVEL] = "VP9E_SET_TARGET_LEVEL", |
|
[VP9E_GET_LEVEL] = "VP9E_GET_LEVEL", |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_ROW_MT |
|
[VP9E_SET_ROW_MT] = "VP9E_SET_ROW_MT", |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TUNE_CONTENT |
|
[VP9E_SET_TUNE_CONTENT] = "VP9E_SET_TUNE_CONTENT", |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TPL |
|
[VP9E_SET_TPL] = "VP9E_SET_TPL", |
|
#endif |
|
#endif |
|
}; |
|
|
|
static av_cold void log_encoder_error(AVCodecContext *avctx, const char *desc) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
const char *error = vpx_codec_error(&ctx->encoder); |
|
const char *detail = vpx_codec_error_detail(&ctx->encoder); |
|
|
|
av_log(avctx, AV_LOG_ERROR, "%s: %s\n", desc, error); |
|
if (detail) |
|
av_log(avctx, AV_LOG_ERROR, " Additional information: %s\n", detail); |
|
} |
|
|
|
static av_cold void dump_enc_cfg(AVCodecContext *avctx, |
|
const struct vpx_codec_enc_cfg *cfg) |
|
{ |
|
int width = -30; |
|
int level = AV_LOG_DEBUG; |
|
int i; |
|
|
|
av_log(avctx, level, "vpx_codec_enc_cfg\n"); |
|
av_log(avctx, level, "generic settings\n" |
|
" %*s%u\n %*s%u\n %*s%u\n %*s%u\n %*s%u\n" |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
" %*s%u\n %*s%u\n" |
|
#endif |
|
" %*s{%u/%u}\n %*s%u\n %*s%d\n %*s%u\n", |
|
width, "g_usage:", cfg->g_usage, |
|
width, "g_threads:", cfg->g_threads, |
|
width, "g_profile:", cfg->g_profile, |
|
width, "g_w:", cfg->g_w, |
|
width, "g_h:", cfg->g_h, |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
width, "g_bit_depth:", cfg->g_bit_depth, |
|
width, "g_input_bit_depth:", cfg->g_input_bit_depth, |
|
#endif |
|
width, "g_timebase:", cfg->g_timebase.num, cfg->g_timebase.den, |
|
width, "g_error_resilient:", cfg->g_error_resilient, |
|
width, "g_pass:", cfg->g_pass, |
|
width, "g_lag_in_frames:", cfg->g_lag_in_frames); |
|
av_log(avctx, level, "rate control settings\n" |
|
" %*s%u\n %*s%u\n %*s%u\n %*s%u\n" |
|
" %*s%d\n %*s%p(%"SIZE_SPECIFIER")\n %*s%u\n", |
|
width, "rc_dropframe_thresh:", cfg->rc_dropframe_thresh, |
|
width, "rc_resize_allowed:", cfg->rc_resize_allowed, |
|
width, "rc_resize_up_thresh:", cfg->rc_resize_up_thresh, |
|
width, "rc_resize_down_thresh:", cfg->rc_resize_down_thresh, |
|
width, "rc_end_usage:", cfg->rc_end_usage, |
|
width, "rc_twopass_stats_in:", cfg->rc_twopass_stats_in.buf, cfg->rc_twopass_stats_in.sz, |
|
width, "rc_target_bitrate:", cfg->rc_target_bitrate); |
|
av_log(avctx, level, "quantizer settings\n" |
|
" %*s%u\n %*s%u\n", |
|
width, "rc_min_quantizer:", cfg->rc_min_quantizer, |
|
width, "rc_max_quantizer:", cfg->rc_max_quantizer); |
|
av_log(avctx, level, "bitrate tolerance\n" |
|
" %*s%u\n %*s%u\n", |
|
width, "rc_undershoot_pct:", cfg->rc_undershoot_pct, |
|
width, "rc_overshoot_pct:", cfg->rc_overshoot_pct); |
|
av_log(avctx, level, "temporal layering settings\n" |
|
" %*s%u\n", width, "ts_number_layers:", cfg->ts_number_layers); |
|
av_log(avctx, level, |
|
"\n %*s", width, "ts_target_bitrate:"); |
|
for (i = 0; i < VPX_TS_MAX_LAYERS; i++) |
|
av_log(avctx, level, "%u ", cfg->ts_target_bitrate[i]); |
|
av_log(avctx, level, "\n"); |
|
av_log(avctx, level, |
|
"\n %*s", width, "ts_rate_decimator:"); |
|
for (i = 0; i < VPX_TS_MAX_LAYERS; i++) |
|
av_log(avctx, level, "%u ", cfg->ts_rate_decimator[i]); |
|
av_log(avctx, level, "\n"); |
|
av_log(avctx, level, |
|
"\n %*s%u\n", width, "ts_periodicity:", cfg->ts_periodicity); |
|
av_log(avctx, level, |
|
"\n %*s", width, "ts_layer_id:"); |
|
for (i = 0; i < VPX_TS_MAX_PERIODICITY; i++) |
|
av_log(avctx, level, "%u ", cfg->ts_layer_id[i]); |
|
av_log(avctx, level, "\n"); |
|
av_log(avctx, level, "decoder buffer model\n" |
|
" %*s%u\n %*s%u\n %*s%u\n", |
|
width, "rc_buf_sz:", cfg->rc_buf_sz, |
|
width, "rc_buf_initial_sz:", cfg->rc_buf_initial_sz, |
|
width, "rc_buf_optimal_sz:", cfg->rc_buf_optimal_sz); |
|
av_log(avctx, level, "2 pass rate control settings\n" |
|
" %*s%u\n %*s%u\n %*s%u\n", |
|
width, "rc_2pass_vbr_bias_pct:", cfg->rc_2pass_vbr_bias_pct, |
|
width, "rc_2pass_vbr_minsection_pct:", cfg->rc_2pass_vbr_minsection_pct, |
|
width, "rc_2pass_vbr_maxsection_pct:", cfg->rc_2pass_vbr_maxsection_pct); |
|
#if VPX_ENCODER_ABI_VERSION >= 14 |
|
av_log(avctx, level, " %*s%u\n", |
|
width, "rc_2pass_vbr_corpus_complexity:", cfg->rc_2pass_vbr_corpus_complexity); |
|
#endif |
|
av_log(avctx, level, "keyframing settings\n" |
|
" %*s%d\n %*s%u\n %*s%u\n", |
|
width, "kf_mode:", cfg->kf_mode, |
|
width, "kf_min_dist:", cfg->kf_min_dist, |
|
width, "kf_max_dist:", cfg->kf_max_dist); |
|
av_log(avctx, level, "\n"); |
|
} |
|
|
|
static void coded_frame_add(void *list, struct FrameListData *cx_frame) |
|
{ |
|
struct FrameListData **p = list; |
|
|
|
while (*p) |
|
p = &(*p)->next; |
|
*p = cx_frame; |
|
cx_frame->next = NULL; |
|
} |
|
|
|
static av_cold void free_coded_frame(struct FrameListData *cx_frame) |
|
{ |
|
av_freep(&cx_frame->buf); |
|
if (cx_frame->buf_alpha) |
|
av_freep(&cx_frame->buf_alpha); |
|
av_freep(&cx_frame); |
|
} |
|
|
|
static av_cold void free_frame_list(struct FrameListData *list) |
|
{ |
|
struct FrameListData *p = list; |
|
|
|
while (p) { |
|
list = list->next; |
|
free_coded_frame(p); |
|
p = list; |
|
} |
|
} |
|
|
|
static av_cold int codecctl_int(AVCodecContext *avctx, |
|
enum vp8e_enc_control_id id, int val) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
char buf[80]; |
|
int width = -30; |
|
int res; |
|
|
|
snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]); |
|
av_log(avctx, AV_LOG_DEBUG, " %*s%d\n", width, buf, val); |
|
|
|
res = vpx_codec_control(&ctx->encoder, id, val); |
|
if (res != VPX_CODEC_OK) { |
|
snprintf(buf, sizeof(buf), "Failed to set %s codec control", |
|
ctlidstr[id]); |
|
log_encoder_error(avctx, buf); |
|
} |
|
|
|
return res == VPX_CODEC_OK ? 0 : AVERROR(EINVAL); |
|
} |
|
|
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
static av_cold int codecctl_intp(AVCodecContext *avctx, |
|
enum vp8e_enc_control_id id, int *val) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
char buf[80]; |
|
int width = -30; |
|
int res; |
|
|
|
snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]); |
|
av_log(avctx, AV_LOG_DEBUG, " %*s%d\n", width, buf, *val); |
|
|
|
res = vpx_codec_control(&ctx->encoder, id, val); |
|
if (res != VPX_CODEC_OK) { |
|
snprintf(buf, sizeof(buf), "Failed to set %s codec control", |
|
ctlidstr[id]); |
|
log_encoder_error(avctx, buf); |
|
} |
|
|
|
return res == VPX_CODEC_OK ? 0 : AVERROR(EINVAL); |
|
} |
|
#endif |
|
|
|
static av_cold int vpx_free(AVCodecContext *avctx) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
|
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
if (avctx->codec_id == AV_CODEC_ID_VP9 && ctx->level >= 0 && |
|
!(avctx->flags & AV_CODEC_FLAG_PASS1)) { |
|
int level_out = 0; |
|
if (!codecctl_intp(avctx, VP9E_GET_LEVEL, &level_out)) |
|
av_log(avctx, AV_LOG_INFO, "Encoded level %.1f\n", level_out * 0.1); |
|
} |
|
#endif |
|
|
|
vpx_codec_destroy(&ctx->encoder); |
|
if (ctx->is_alpha) { |
|
vpx_codec_destroy(&ctx->encoder_alpha); |
|
av_freep(&ctx->rawimg_alpha.planes[VPX_PLANE_U]); |
|
av_freep(&ctx->rawimg_alpha.planes[VPX_PLANE_V]); |
|
} |
|
av_freep(&ctx->twopass_stats.buf); |
|
av_freep(&avctx->stats_out); |
|
free_frame_list(ctx->coded_frame_list); |
|
return 0; |
|
} |
|
|
|
static void vp8_ts_parse_int_array(int *dest, char *value, size_t value_len, int max_entries) |
|
{ |
|
int dest_idx = 0; |
|
char *saveptr = NULL; |
|
char *token = av_strtok(value, ",", &saveptr); |
|
|
|
while (token && dest_idx < max_entries) { |
|
dest[dest_idx++] = strtoul(token, NULL, 10); |
|
token = av_strtok(NULL, ",", &saveptr); |
|
} |
|
} |
|
|
|
static int vp8_ts_param_parse(struct vpx_codec_enc_cfg *enccfg, char *key, char *value) |
|
{ |
|
size_t value_len = strlen(value); |
|
|
|
if (!value_len) |
|
return -1; |
|
|
|
if (!strcmp(key, "ts_number_layers")) |
|
enccfg->ts_number_layers = strtoul(value, &value, 10); |
|
else if (!strcmp(key, "ts_target_bitrate")) |
|
vp8_ts_parse_int_array(enccfg->ts_target_bitrate, value, value_len, VPX_TS_MAX_LAYERS); |
|
else if (!strcmp(key, "ts_rate_decimator")) |
|
vp8_ts_parse_int_array(enccfg->ts_rate_decimator, value, value_len, VPX_TS_MAX_LAYERS); |
|
else if (!strcmp(key, "ts_periodicity")) |
|
enccfg->ts_periodicity = strtoul(value, &value, 10); |
|
else if (!strcmp(key, "ts_layer_id")) |
|
vp8_ts_parse_int_array(enccfg->ts_layer_id, value, value_len, VPX_TS_MAX_PERIODICITY); |
|
|
|
return 0; |
|
} |
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
static int set_pix_fmt(AVCodecContext *avctx, vpx_codec_caps_t codec_caps, |
|
struct vpx_codec_enc_cfg *enccfg, vpx_codec_flags_t *flags, |
|
vpx_img_fmt_t *img_fmt) |
|
{ |
|
VPxContext av_unused *ctx = avctx->priv_data; |
|
enccfg->g_bit_depth = enccfg->g_input_bit_depth = 8; |
|
switch (avctx->pix_fmt) { |
|
case AV_PIX_FMT_YUV420P: |
|
case AV_PIX_FMT_YUVA420P: |
|
enccfg->g_profile = 0; |
|
*img_fmt = VPX_IMG_FMT_I420; |
|
return 0; |
|
case AV_PIX_FMT_YUV422P: |
|
enccfg->g_profile = 1; |
|
*img_fmt = VPX_IMG_FMT_I422; |
|
return 0; |
|
case AV_PIX_FMT_YUV440P: |
|
enccfg->g_profile = 1; |
|
*img_fmt = VPX_IMG_FMT_I440; |
|
return 0; |
|
case AV_PIX_FMT_GBRP: |
|
ctx->vpx_cs = VPX_CS_SRGB; |
|
case AV_PIX_FMT_YUV444P: |
|
enccfg->g_profile = 1; |
|
*img_fmt = VPX_IMG_FMT_I444; |
|
return 0; |
|
case AV_PIX_FMT_YUV420P10: |
|
case AV_PIX_FMT_YUV420P12: |
|
if (codec_caps & VPX_CODEC_CAP_HIGHBITDEPTH) { |
|
enccfg->g_bit_depth = enccfg->g_input_bit_depth = |
|
avctx->pix_fmt == AV_PIX_FMT_YUV420P10 ? 10 : 12; |
|
enccfg->g_profile = 2; |
|
*img_fmt = VPX_IMG_FMT_I42016; |
|
*flags |= VPX_CODEC_USE_HIGHBITDEPTH; |
|
return 0; |
|
} |
|
break; |
|
case AV_PIX_FMT_YUV422P10: |
|
case AV_PIX_FMT_YUV422P12: |
|
if (codec_caps & VPX_CODEC_CAP_HIGHBITDEPTH) { |
|
enccfg->g_bit_depth = enccfg->g_input_bit_depth = |
|
avctx->pix_fmt == AV_PIX_FMT_YUV422P10 ? 10 : 12; |
|
enccfg->g_profile = 3; |
|
*img_fmt = VPX_IMG_FMT_I42216; |
|
*flags |= VPX_CODEC_USE_HIGHBITDEPTH; |
|
return 0; |
|
} |
|
break; |
|
case AV_PIX_FMT_YUV440P10: |
|
case AV_PIX_FMT_YUV440P12: |
|
if (codec_caps & VPX_CODEC_CAP_HIGHBITDEPTH) { |
|
enccfg->g_bit_depth = enccfg->g_input_bit_depth = |
|
avctx->pix_fmt == AV_PIX_FMT_YUV440P10 ? 10 : 12; |
|
enccfg->g_profile = 3; |
|
*img_fmt = VPX_IMG_FMT_I44016; |
|
*flags |= VPX_CODEC_USE_HIGHBITDEPTH; |
|
return 0; |
|
} |
|
break; |
|
case AV_PIX_FMT_GBRP10: |
|
case AV_PIX_FMT_GBRP12: |
|
ctx->vpx_cs = VPX_CS_SRGB; |
|
case AV_PIX_FMT_YUV444P10: |
|
case AV_PIX_FMT_YUV444P12: |
|
if (codec_caps & VPX_CODEC_CAP_HIGHBITDEPTH) { |
|
enccfg->g_bit_depth = enccfg->g_input_bit_depth = |
|
avctx->pix_fmt == AV_PIX_FMT_YUV444P10 || |
|
avctx->pix_fmt == AV_PIX_FMT_GBRP10 ? 10 : 12; |
|
enccfg->g_profile = 3; |
|
*img_fmt = VPX_IMG_FMT_I44416; |
|
*flags |= VPX_CODEC_USE_HIGHBITDEPTH; |
|
return 0; |
|
} |
|
break; |
|
default: |
|
break; |
|
} |
|
av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format.\n"); |
|
return AVERROR_INVALIDDATA; |
|
} |
|
|
|
static void set_colorspace(AVCodecContext *avctx) |
|
{ |
|
enum vpx_color_space vpx_cs; |
|
VPxContext *ctx = avctx->priv_data; |
|
|
|
if (ctx->vpx_cs) { |
|
vpx_cs = ctx->vpx_cs; |
|
} else { |
|
switch (avctx->colorspace) { |
|
case AVCOL_SPC_RGB: vpx_cs = VPX_CS_SRGB; break; |
|
case AVCOL_SPC_BT709: vpx_cs = VPX_CS_BT_709; break; |
|
case AVCOL_SPC_UNSPECIFIED: vpx_cs = VPX_CS_UNKNOWN; break; |
|
case AVCOL_SPC_RESERVED: vpx_cs = VPX_CS_RESERVED; break; |
|
case AVCOL_SPC_BT470BG: vpx_cs = VPX_CS_BT_601; break; |
|
case AVCOL_SPC_SMPTE170M: vpx_cs = VPX_CS_SMPTE_170; break; |
|
case AVCOL_SPC_SMPTE240M: vpx_cs = VPX_CS_SMPTE_240; break; |
|
case AVCOL_SPC_BT2020_NCL: vpx_cs = VPX_CS_BT_2020; break; |
|
default: |
|
av_log(avctx, AV_LOG_WARNING, "Unsupported colorspace (%d)\n", |
|
avctx->colorspace); |
|
return; |
|
} |
|
} |
|
codecctl_int(avctx, VP9E_SET_COLOR_SPACE, vpx_cs); |
|
} |
|
|
|
#if VPX_ENCODER_ABI_VERSION >= 11 |
|
static void set_color_range(AVCodecContext *avctx) |
|
{ |
|
enum vpx_color_range vpx_cr; |
|
switch (avctx->color_range) { |
|
case AVCOL_RANGE_UNSPECIFIED: |
|
case AVCOL_RANGE_MPEG: vpx_cr = VPX_CR_STUDIO_RANGE; break; |
|
case AVCOL_RANGE_JPEG: vpx_cr = VPX_CR_FULL_RANGE; break; |
|
default: |
|
av_log(avctx, AV_LOG_WARNING, "Unsupported color range (%d)\n", |
|
avctx->color_range); |
|
return; |
|
} |
|
|
|
codecctl_int(avctx, VP9E_SET_COLOR_RANGE, vpx_cr); |
|
} |
|
#endif |
|
#endif |
|
|
|
/** |
|
* Set the target bitrate to VPX library default. Also set CRF to 32 if needed. |
|
*/ |
|
static void set_vp8_defaults(AVCodecContext *avctx, |
|
struct vpx_codec_enc_cfg *enccfg) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
av_assert0(!avctx->bit_rate); |
|
avctx->bit_rate = enccfg->rc_target_bitrate * 1000; |
|
if (enccfg->rc_end_usage == VPX_CQ) { |
|
av_log(avctx, AV_LOG_WARNING, |
|
"Bitrate not specified for constrained quality mode, using default of %dkbit/sec\n", |
|
enccfg->rc_target_bitrate); |
|
} else { |
|
enccfg->rc_end_usage = VPX_CQ; |
|
ctx->crf = 32; |
|
av_log(avctx, AV_LOG_WARNING, |
|
"Neither bitrate nor constrained quality specified, using default CRF of %d and bitrate of %dkbit/sec\n", |
|
ctx->crf, enccfg->rc_target_bitrate); |
|
} |
|
} |
|
|
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
/** |
|
* Keep the target bitrate at 0 to engage constant quality mode. If CRF is not |
|
* set, use 32. |
|
*/ |
|
static void set_vp9_defaults(AVCodecContext *avctx, |
|
struct vpx_codec_enc_cfg *enccfg) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
av_assert0(!avctx->bit_rate); |
|
if (enccfg->rc_end_usage != VPX_Q && ctx->lossless < 0) { |
|
enccfg->rc_end_usage = VPX_Q; |
|
ctx->crf = 32; |
|
av_log(avctx, AV_LOG_WARNING, |
|
"Neither bitrate nor constrained quality specified, using default CRF of %d\n", |
|
ctx->crf); |
|
} |
|
} |
|
#endif |
|
|
|
/** |
|
* Called when the bitrate is not set. It sets appropriate default values for |
|
* bitrate and CRF. |
|
*/ |
|
static void set_vpx_defaults(AVCodecContext *avctx, |
|
struct vpx_codec_enc_cfg *enccfg) |
|
{ |
|
av_assert0(!avctx->bit_rate); |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (avctx->codec_id == AV_CODEC_ID_VP9) { |
|
set_vp9_defaults(avctx, enccfg); |
|
return; |
|
} |
|
#endif |
|
set_vp8_defaults(avctx, enccfg); |
|
} |
|
|
|
static av_cold int vpx_init(AVCodecContext *avctx, |
|
const struct vpx_codec_iface *iface) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
struct vpx_codec_enc_cfg enccfg = { 0 }; |
|
struct vpx_codec_enc_cfg enccfg_alpha; |
|
vpx_codec_flags_t flags = (avctx->flags & AV_CODEC_FLAG_PSNR) ? VPX_CODEC_USE_PSNR : 0; |
|
AVCPBProperties *cpb_props; |
|
int res; |
|
vpx_img_fmt_t img_fmt = VPX_IMG_FMT_I420; |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
vpx_codec_caps_t codec_caps = vpx_codec_get_caps(iface); |
|
#endif |
|
|
|
av_log(avctx, AV_LOG_INFO, "%s\n", vpx_codec_version_str()); |
|
av_log(avctx, AV_LOG_VERBOSE, "%s\n", vpx_codec_build_config()); |
|
|
|
if (avctx->pix_fmt == AV_PIX_FMT_YUVA420P) |
|
ctx->is_alpha = 1; |
|
|
|
if ((res = vpx_codec_enc_config_default(iface, &enccfg, 0)) != VPX_CODEC_OK) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to get config: %s\n", |
|
vpx_codec_err_to_string(res)); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (avctx->codec_id == AV_CODEC_ID_VP9) { |
|
if (set_pix_fmt(avctx, codec_caps, &enccfg, &flags, &img_fmt)) |
|
return AVERROR(EINVAL); |
|
} |
|
#endif |
|
|
|
if(!avctx->bit_rate) |
|
if(avctx->rc_max_rate || avctx->rc_buffer_size || avctx->rc_initial_buffer_occupancy) { |
|
av_log( avctx, AV_LOG_ERROR, "Rate control parameters set without a bitrate\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
dump_enc_cfg(avctx, &enccfg); |
|
|
|
enccfg.g_w = avctx->width; |
|
enccfg.g_h = avctx->height; |
|
enccfg.g_timebase.num = avctx->time_base.num; |
|
enccfg.g_timebase.den = avctx->time_base.den; |
|
enccfg.g_threads = |
|
FFMIN(avctx->thread_count ? avctx->thread_count : av_cpu_count(), 16); |
|
enccfg.g_lag_in_frames= ctx->lag_in_frames; |
|
|
|
if (avctx->flags & AV_CODEC_FLAG_PASS1) |
|
enccfg.g_pass = VPX_RC_FIRST_PASS; |
|
else if (avctx->flags & AV_CODEC_FLAG_PASS2) |
|
enccfg.g_pass = VPX_RC_LAST_PASS; |
|
else |
|
enccfg.g_pass = VPX_RC_ONE_PASS; |
|
|
|
if (avctx->rc_min_rate == avctx->rc_max_rate && |
|
avctx->rc_min_rate == avctx->bit_rate && avctx->bit_rate) { |
|
enccfg.rc_end_usage = VPX_CBR; |
|
} else if (ctx->crf >= 0) { |
|
enccfg.rc_end_usage = VPX_CQ; |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (!avctx->bit_rate && avctx->codec_id == AV_CODEC_ID_VP9) |
|
enccfg.rc_end_usage = VPX_Q; |
|
#endif |
|
} |
|
|
|
if (avctx->bit_rate) { |
|
enccfg.rc_target_bitrate = av_rescale_rnd(avctx->bit_rate, 1, 1000, |
|
AV_ROUND_NEAR_INF); |
|
} else { |
|
// Set bitrate to default value. Also sets CRF to default if needed. |
|
set_vpx_defaults(avctx, &enccfg); |
|
} |
|
|
|
if (avctx->codec_id == AV_CODEC_ID_VP9 && ctx->lossless == 1) { |
|
enccfg.rc_min_quantizer = |
|
enccfg.rc_max_quantizer = 0; |
|
} else { |
|
if (avctx->qmin >= 0) |
|
enccfg.rc_min_quantizer = avctx->qmin; |
|
if (avctx->qmax >= 0) |
|
enccfg.rc_max_quantizer = avctx->qmax; |
|
} |
|
|
|
if (enccfg.rc_end_usage == VPX_CQ |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
|| enccfg.rc_end_usage == VPX_Q |
|
#endif |
|
) { |
|
if (ctx->crf < enccfg.rc_min_quantizer || ctx->crf > enccfg.rc_max_quantizer) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"CQ level %d must be between minimum and maximum quantizer value (%d-%d)\n", |
|
ctx->crf, enccfg.rc_min_quantizer, enccfg.rc_max_quantizer); |
|
return AVERROR(EINVAL); |
|
} |
|
} |
|
|
|
#if FF_API_PRIVATE_OPT |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
if (avctx->frame_skip_threshold) |
|
ctx->drop_threshold = avctx->frame_skip_threshold; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
enccfg.rc_dropframe_thresh = ctx->drop_threshold; |
|
|
|
//0-100 (0 => CBR, 100 => VBR) |
|
enccfg.rc_2pass_vbr_bias_pct = lrint(avctx->qcompress * 100); |
|
if (avctx->bit_rate) |
|
enccfg.rc_2pass_vbr_minsection_pct = |
|
avctx->rc_min_rate * 100LL / avctx->bit_rate; |
|
if (avctx->rc_max_rate) |
|
enccfg.rc_2pass_vbr_maxsection_pct = |
|
avctx->rc_max_rate * 100LL / avctx->bit_rate; |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (avctx->codec_id == AV_CODEC_ID_VP9) { |
|
#if VPX_ENCODER_ABI_VERSION >= 14 |
|
if (ctx->corpus_complexity >= 0) |
|
enccfg.rc_2pass_vbr_corpus_complexity = ctx->corpus_complexity; |
|
#endif |
|
} |
|
#endif |
|
|
|
if (avctx->rc_buffer_size) |
|
enccfg.rc_buf_sz = |
|
avctx->rc_buffer_size * 1000LL / avctx->bit_rate; |
|
if (avctx->rc_initial_buffer_occupancy) |
|
enccfg.rc_buf_initial_sz = |
|
avctx->rc_initial_buffer_occupancy * 1000LL / avctx->bit_rate; |
|
enccfg.rc_buf_optimal_sz = enccfg.rc_buf_sz * 5 / 6; |
|
if (ctx->rc_undershoot_pct >= 0) |
|
enccfg.rc_undershoot_pct = ctx->rc_undershoot_pct; |
|
if (ctx->rc_overshoot_pct >= 0) |
|
enccfg.rc_overshoot_pct = ctx->rc_overshoot_pct; |
|
|
|
//_enc_init() will balk if kf_min_dist differs from max w/VPX_KF_AUTO |
|
if (avctx->keyint_min >= 0 && avctx->keyint_min == avctx->gop_size) |
|
enccfg.kf_min_dist = avctx->keyint_min; |
|
if (avctx->gop_size >= 0) |
|
enccfg.kf_max_dist = avctx->gop_size; |
|
|
|
if (enccfg.g_pass == VPX_RC_FIRST_PASS) |
|
enccfg.g_lag_in_frames = 0; |
|
else if (enccfg.g_pass == VPX_RC_LAST_PASS) { |
|
int decode_size, ret; |
|
|
|
if (!avctx->stats_in) { |
|
av_log(avctx, AV_LOG_ERROR, "No stats file for second pass\n"); |
|
return AVERROR_INVALIDDATA; |
|
} |
|
|
|
ctx->twopass_stats.sz = strlen(avctx->stats_in) * 3 / 4; |
|
ret = av_reallocp(&ctx->twopass_stats.buf, ctx->twopass_stats.sz); |
|
if (ret < 0) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Stat buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n", |
|
ctx->twopass_stats.sz); |
|
ctx->twopass_stats.sz = 0; |
|
return ret; |
|
} |
|
decode_size = av_base64_decode(ctx->twopass_stats.buf, avctx->stats_in, |
|
ctx->twopass_stats.sz); |
|
if (decode_size < 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Stat buffer decode failed\n"); |
|
return AVERROR_INVALIDDATA; |
|
} |
|
|
|
ctx->twopass_stats.sz = decode_size; |
|
enccfg.rc_twopass_stats_in = ctx->twopass_stats; |
|
} |
|
|
|
/* 0-3: For non-zero values the encoder increasingly optimizes for reduced |
|
complexity playback on low powered devices at the expense of encode |
|
quality. */ |
|
if (avctx->profile != FF_PROFILE_UNKNOWN) |
|
enccfg.g_profile = avctx->profile; |
|
|
|
enccfg.g_error_resilient = ctx->error_resilient || ctx->flags & VP8F_ERROR_RESILIENT; |
|
|
|
if (CONFIG_LIBVPX_VP8_ENCODER && avctx->codec_id == AV_CODEC_ID_VP8) { |
|
AVDictionaryEntry* en = NULL; |
|
while ((en = av_dict_get(ctx->vp8_ts_parameters, "", en, AV_DICT_IGNORE_SUFFIX))) { |
|
if (vp8_ts_param_parse(&enccfg, en->key, en->value) < 0) |
|
av_log(avctx, AV_LOG_WARNING, |
|
"Error parsing option '%s = %s'.\n", |
|
en->key, en->value); |
|
} |
|
} |
|
|
|
dump_enc_cfg(avctx, &enccfg); |
|
/* Construct Encoder Context */ |
|
res = vpx_codec_enc_init(&ctx->encoder, iface, &enccfg, flags); |
|
if (res != VPX_CODEC_OK) { |
|
log_encoder_error(avctx, "Failed to initialize encoder"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (ctx->is_alpha) { |
|
enccfg_alpha = enccfg; |
|
res = vpx_codec_enc_init(&ctx->encoder_alpha, iface, &enccfg_alpha, flags); |
|
if (res != VPX_CODEC_OK) { |
|
log_encoder_error(avctx, "Failed to initialize alpha encoder"); |
|
return AVERROR(EINVAL); |
|
} |
|
} |
|
|
|
//codec control failures are currently treated only as warnings |
|
av_log(avctx, AV_LOG_DEBUG, "vpx_codec_control\n"); |
|
codecctl_int(avctx, VP8E_SET_CPUUSED, ctx->cpu_used); |
|
if (ctx->flags & VP8F_AUTO_ALT_REF) |
|
ctx->auto_alt_ref = 1; |
|
if (ctx->auto_alt_ref >= 0) |
|
codecctl_int(avctx, VP8E_SET_ENABLEAUTOALTREF, |
|
avctx->codec_id == AV_CODEC_ID_VP8 ? !!ctx->auto_alt_ref : ctx->auto_alt_ref); |
|
if (ctx->arnr_max_frames >= 0) |
|
codecctl_int(avctx, VP8E_SET_ARNR_MAXFRAMES, ctx->arnr_max_frames); |
|
if (ctx->arnr_strength >= 0) |
|
codecctl_int(avctx, VP8E_SET_ARNR_STRENGTH, ctx->arnr_strength); |
|
if (ctx->arnr_type >= 0) |
|
codecctl_int(avctx, VP8E_SET_ARNR_TYPE, ctx->arnr_type); |
|
if (ctx->tune >= 0) |
|
codecctl_int(avctx, VP8E_SET_TUNING, ctx->tune); |
|
|
|
if (ctx->auto_alt_ref && ctx->is_alpha && avctx->codec_id == AV_CODEC_ID_VP8) { |
|
av_log(avctx, AV_LOG_ERROR, "Transparency encoding with auto_alt_ref does not work\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (ctx->sharpness >= 0) |
|
codecctl_int(avctx, VP8E_SET_SHARPNESS, ctx->sharpness); |
|
|
|
if (CONFIG_LIBVPX_VP8_ENCODER && avctx->codec_id == AV_CODEC_ID_VP8) { |
|
#if FF_API_PRIVATE_OPT |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
if (avctx->noise_reduction) |
|
ctx->noise_sensitivity = avctx->noise_reduction; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
codecctl_int(avctx, VP8E_SET_NOISE_SENSITIVITY, ctx->noise_sensitivity); |
|
codecctl_int(avctx, VP8E_SET_TOKEN_PARTITIONS, av_log2(avctx->slices)); |
|
} |
|
codecctl_int(avctx, VP8E_SET_STATIC_THRESHOLD, ctx->static_thresh); |
|
if (ctx->crf >= 0) |
|
codecctl_int(avctx, VP8E_SET_CQ_LEVEL, ctx->crf); |
|
if (ctx->max_intra_rate >= 0) |
|
codecctl_int(avctx, VP8E_SET_MAX_INTRA_BITRATE_PCT, ctx->max_intra_rate); |
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (avctx->codec_id == AV_CODEC_ID_VP9) { |
|
if (ctx->lossless >= 0) |
|
codecctl_int(avctx, VP9E_SET_LOSSLESS, ctx->lossless); |
|
if (ctx->tile_columns >= 0) |
|
codecctl_int(avctx, VP9E_SET_TILE_COLUMNS, ctx->tile_columns); |
|
if (ctx->tile_rows >= 0) |
|
codecctl_int(avctx, VP9E_SET_TILE_ROWS, ctx->tile_rows); |
|
if (ctx->frame_parallel >= 0) |
|
codecctl_int(avctx, VP9E_SET_FRAME_PARALLEL_DECODING, ctx->frame_parallel); |
|
if (ctx->aq_mode >= 0) |
|
codecctl_int(avctx, VP9E_SET_AQ_MODE, ctx->aq_mode); |
|
set_colorspace(avctx); |
|
#if VPX_ENCODER_ABI_VERSION >= 11 |
|
set_color_range(avctx); |
|
#endif |
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
codecctl_int(avctx, VP9E_SET_TARGET_LEVEL, ctx->level < 0 ? 255 : lrint(ctx->level * 10)); |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_ROW_MT |
|
if (ctx->row_mt >= 0) |
|
codecctl_int(avctx, VP9E_SET_ROW_MT, ctx->row_mt); |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TUNE_CONTENT |
|
if (ctx->tune_content >= 0) |
|
codecctl_int(avctx, VP9E_SET_TUNE_CONTENT, ctx->tune_content); |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TPL |
|
if (ctx->tpl_model >= 0) |
|
codecctl_int(avctx, VP9E_SET_TPL, ctx->tpl_model); |
|
#endif |
|
} |
|
#endif |
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Using deadline: %d\n", ctx->deadline); |
|
|
|
//provide dummy value to initialize wrapper, values will be updated each _encode() |
|
vpx_img_wrap(&ctx->rawimg, img_fmt, avctx->width, avctx->height, 1, |
|
(unsigned char*)1); |
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
if (avctx->codec_id == AV_CODEC_ID_VP9 && (codec_caps & VPX_CODEC_CAP_HIGHBITDEPTH)) |
|
ctx->rawimg.bit_depth = enccfg.g_bit_depth; |
|
#endif |
|
|
|
cpb_props = ff_add_cpb_side_data(avctx); |
|
if (!cpb_props) |
|
return AVERROR(ENOMEM); |
|
|
|
if (enccfg.rc_end_usage == VPX_CBR || |
|
enccfg.g_pass != VPX_RC_ONE_PASS) { |
|
cpb_props->max_bitrate = avctx->rc_max_rate; |
|
cpb_props->min_bitrate = avctx->rc_min_rate; |
|
cpb_props->avg_bitrate = avctx->bit_rate; |
|
} |
|
cpb_props->buffer_size = avctx->rc_buffer_size; |
|
|
|
return 0; |
|
} |
|
|
|
static inline void cx_pktcpy(struct FrameListData *dst, |
|
const struct vpx_codec_cx_pkt *src, |
|
const struct vpx_codec_cx_pkt *src_alpha, |
|
VPxContext *ctx) |
|
{ |
|
dst->pts = src->data.frame.pts; |
|
dst->duration = src->data.frame.duration; |
|
dst->flags = src->data.frame.flags; |
|
dst->sz = src->data.frame.sz; |
|
dst->buf = src->data.frame.buf; |
|
dst->have_sse = 0; |
|
/* For alt-ref frame, don't store PSNR or increment frame_number */ |
|
if (!(dst->flags & VPX_FRAME_IS_INVISIBLE)) { |
|
dst->frame_number = ++ctx->frame_number; |
|
dst->have_sse = ctx->have_sse; |
|
if (ctx->have_sse) { |
|
/* associate last-seen SSE to the frame. */ |
|
/* Transfers ownership from ctx to dst. */ |
|
/* WARNING! This makes the assumption that PSNR_PKT comes |
|
just before the frame it refers to! */ |
|
memcpy(dst->sse, ctx->sse, sizeof(dst->sse)); |
|
ctx->have_sse = 0; |
|
} |
|
} else { |
|
dst->frame_number = -1; /* sanity marker */ |
|
} |
|
if (src_alpha) { |
|
dst->buf_alpha = src_alpha->data.frame.buf; |
|
dst->sz_alpha = src_alpha->data.frame.sz; |
|
} else { |
|
dst->buf_alpha = NULL; |
|
dst->sz_alpha = 0; |
|
} |
|
} |
|
|
|
/** |
|
* Store coded frame information in format suitable for return from encode2(). |
|
* |
|
* Write information from @a cx_frame to @a pkt |
|
* @return packet data size on success |
|
* @return a negative AVERROR on error |
|
*/ |
|
static int storeframe(AVCodecContext *avctx, struct FrameListData *cx_frame, |
|
AVPacket *pkt) |
|
{ |
|
int ret = ff_alloc_packet2(avctx, pkt, cx_frame->sz, 0); |
|
uint8_t *side_data; |
|
if (ret >= 0) { |
|
int pict_type; |
|
memcpy(pkt->data, cx_frame->buf, pkt->size); |
|
pkt->pts = pkt->dts = cx_frame->pts; |
|
#if FF_API_CODED_FRAME |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
avctx->coded_frame->pts = cx_frame->pts; |
|
avctx->coded_frame->key_frame = !!(cx_frame->flags & VPX_FRAME_IS_KEY); |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
|
|
if (!!(cx_frame->flags & VPX_FRAME_IS_KEY)) { |
|
pict_type = AV_PICTURE_TYPE_I; |
|
#if FF_API_CODED_FRAME |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
avctx->coded_frame->pict_type = pict_type; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
pkt->flags |= AV_PKT_FLAG_KEY; |
|
} else { |
|
pict_type = AV_PICTURE_TYPE_P; |
|
#if FF_API_CODED_FRAME |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
avctx->coded_frame->pict_type = pict_type; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
} |
|
|
|
ff_side_data_set_encoder_stats(pkt, 0, cx_frame->sse + 1, |
|
cx_frame->have_sse ? 3 : 0, pict_type); |
|
|
|
if (cx_frame->have_sse) { |
|
int i; |
|
/* Beware of the Y/U/V/all order! */ |
|
#if FF_API_CODED_FRAME |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
avctx->coded_frame->error[0] = cx_frame->sse[1]; |
|
avctx->coded_frame->error[1] = cx_frame->sse[2]; |
|
avctx->coded_frame->error[2] = cx_frame->sse[3]; |
|
avctx->coded_frame->error[3] = 0; // alpha |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
for (i = 0; i < 3; ++i) { |
|
avctx->error[i] += cx_frame->sse[i + 1]; |
|
} |
|
cx_frame->have_sse = 0; |
|
} |
|
if (cx_frame->sz_alpha > 0) { |
|
side_data = av_packet_new_side_data(pkt, |
|
AV_PKT_DATA_MATROSKA_BLOCKADDITIONAL, |
|
cx_frame->sz_alpha + 8); |
|
if(!side_data) { |
|
av_packet_unref(pkt); |
|
av_free(pkt); |
|
return AVERROR(ENOMEM); |
|
} |
|
AV_WB64(side_data, 1); |
|
memcpy(side_data + 8, cx_frame->buf_alpha, cx_frame->sz_alpha); |
|
} |
|
} else { |
|
return ret; |
|
} |
|
return pkt->size; |
|
} |
|
|
|
/** |
|
* Queue multiple output frames from the encoder, returning the front-most. |
|
* In cases where vpx_codec_get_cx_data() returns more than 1 frame append |
|
* the frame queue. Return the head frame if available. |
|
* @return Stored frame size |
|
* @return AVERROR(EINVAL) on output size error |
|
* @return AVERROR(ENOMEM) on coded frame queue data allocation error |
|
*/ |
|
static int queue_frames(AVCodecContext *avctx, AVPacket *pkt_out) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
const struct vpx_codec_cx_pkt *pkt; |
|
const struct vpx_codec_cx_pkt *pkt_alpha = NULL; |
|
const void *iter = NULL; |
|
const void *iter_alpha = NULL; |
|
int size = 0; |
|
|
|
if (ctx->coded_frame_list) { |
|
struct FrameListData *cx_frame = ctx->coded_frame_list; |
|
/* return the leading frame if we've already begun queueing */ |
|
size = storeframe(avctx, cx_frame, pkt_out); |
|
if (size < 0) |
|
return size; |
|
ctx->coded_frame_list = cx_frame->next; |
|
free_coded_frame(cx_frame); |
|
} |
|
|
|
/* consume all available output from the encoder before returning. buffers |
|
are only good through the next vpx_codec call */ |
|
while ((pkt = vpx_codec_get_cx_data(&ctx->encoder, &iter)) && |
|
(!ctx->is_alpha || |
|
(pkt_alpha = vpx_codec_get_cx_data(&ctx->encoder_alpha, &iter_alpha)))) { |
|
switch (pkt->kind) { |
|
case VPX_CODEC_CX_FRAME_PKT: |
|
if (!size) { |
|
struct FrameListData cx_frame; |
|
|
|
/* avoid storing the frame when the list is empty and we haven't yet |
|
provided a frame for output */ |
|
av_assert0(!ctx->coded_frame_list); |
|
cx_pktcpy(&cx_frame, pkt, pkt_alpha, ctx); |
|
size = storeframe(avctx, &cx_frame, pkt_out); |
|
if (size < 0) |
|
return size; |
|
} else { |
|
struct FrameListData *cx_frame = av_malloc(sizeof(*cx_frame)); |
|
|
|
if (!cx_frame) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Frame queue element alloc failed\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
cx_pktcpy(cx_frame, pkt, pkt_alpha, ctx); |
|
cx_frame->buf = av_malloc(cx_frame->sz); |
|
|
|
if (!cx_frame->buf) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Data buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n", |
|
cx_frame->sz); |
|
av_freep(&cx_frame); |
|
return AVERROR(ENOMEM); |
|
} |
|
memcpy(cx_frame->buf, pkt->data.frame.buf, pkt->data.frame.sz); |
|
if (ctx->is_alpha) { |
|
cx_frame->buf_alpha = av_malloc(cx_frame->sz_alpha); |
|
if (!cx_frame->buf_alpha) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Data buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n", |
|
cx_frame->sz_alpha); |
|
av_free(cx_frame); |
|
return AVERROR(ENOMEM); |
|
} |
|
memcpy(cx_frame->buf_alpha, pkt_alpha->data.frame.buf, pkt_alpha->data.frame.sz); |
|
} |
|
coded_frame_add(&ctx->coded_frame_list, cx_frame); |
|
} |
|
break; |
|
case VPX_CODEC_STATS_PKT: { |
|
struct vpx_fixed_buf *stats = &ctx->twopass_stats; |
|
int err; |
|
if ((err = av_reallocp(&stats->buf, |
|
stats->sz + |
|
pkt->data.twopass_stats.sz)) < 0) { |
|
stats->sz = 0; |
|
av_log(avctx, AV_LOG_ERROR, "Stat buffer realloc failed\n"); |
|
return err; |
|
} |
|
memcpy((uint8_t*)stats->buf + stats->sz, |
|
pkt->data.twopass_stats.buf, pkt->data.twopass_stats.sz); |
|
stats->sz += pkt->data.twopass_stats.sz; |
|
break; |
|
} |
|
case VPX_CODEC_PSNR_PKT: |
|
av_assert0(!ctx->have_sse); |
|
ctx->sse[0] = pkt->data.psnr.sse[0]; |
|
ctx->sse[1] = pkt->data.psnr.sse[1]; |
|
ctx->sse[2] = pkt->data.psnr.sse[2]; |
|
ctx->sse[3] = pkt->data.psnr.sse[3]; |
|
ctx->have_sse = 1; |
|
break; |
|
case VPX_CODEC_CUSTOM_PKT: |
|
//ignore unsupported/unrecognized packet types |
|
break; |
|
} |
|
} |
|
|
|
return size; |
|
} |
|
|
|
static int set_roi_map(AVCodecContext *avctx, const AVFrameSideData *sd, int frame_width, int frame_height, |
|
vpx_roi_map_t *roi_map, int block_size, int segment_cnt) |
|
{ |
|
/** |
|
* range of vpx_roi_map_t.delta_q[i] is [-63, 63] |
|
*/ |
|
#define MAX_DELTA_Q 63 |
|
|
|
const AVRegionOfInterest *roi = NULL; |
|
int nb_rois; |
|
uint32_t self_size; |
|
int segment_id; |
|
|
|
/* record the mapping from delta_q to "segment id + 1" in segment_mapping[]. |
|
* the range of delta_q is [-MAX_DELTA_Q, MAX_DELTA_Q], |
|
* and its corresponding array index is [0, 2 * MAX_DELTA_Q], |
|
* and so the length of the mapping array is 2 * MAX_DELTA_Q + 1. |
|
* "segment id + 1", so we can say there's no mapping if the value of array element is zero. |
|
*/ |
|
int segment_mapping[2 * MAX_DELTA_Q + 1] = { 0 }; |
|
|
|
memset(roi_map, 0, sizeof(*roi_map)); |
|
|
|
/* segment id 0 in roi_map is reserved for the areas not covered by AVRegionOfInterest. |
|
* segment id 0 in roi_map is also for the areas with AVRegionOfInterest.qoffset near 0. |
|
* (delta_q of segment id 0 is 0). |
|
*/ |
|
segment_mapping[MAX_DELTA_Q] = 1; |
|
segment_id = 1; |
|
|
|
roi = (const AVRegionOfInterest*)sd->data; |
|
self_size = roi->self_size; |
|
if (!self_size || sd->size % self_size) { |
|
av_log(avctx, AV_LOG_ERROR, "Invalid AVRegionOfInterest.self_size.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
nb_rois = sd->size / self_size; |
|
|
|
/* This list must be iterated from zero because regions are |
|
* defined in order of decreasing importance. So discard less |
|
* important areas if they exceed the segment count. |
|
*/ |
|
for (int i = 0; i < nb_rois; i++) { |
|
int delta_q; |
|
int mapping_index; |
|
|
|
roi = (const AVRegionOfInterest*)(sd->data + self_size * i); |
|
if (!roi->qoffset.den) { |
|
av_log(avctx, AV_LOG_ERROR, "AVRegionOfInterest.qoffset.den must not be zero.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
delta_q = (int)(roi->qoffset.num * 1.0f / roi->qoffset.den * MAX_DELTA_Q); |
|
delta_q = av_clip(delta_q, -MAX_DELTA_Q, MAX_DELTA_Q); |
|
|
|
mapping_index = delta_q + MAX_DELTA_Q; |
|
if (!segment_mapping[mapping_index]) { |
|
if (segment_id == segment_cnt) { |
|
av_log(avctx, AV_LOG_WARNING, |
|
"ROI only supports %d segments (and segment 0 is reserved for non-ROIs), skipping the left ones.\n", |
|
segment_cnt); |
|
break; |
|
} |
|
|
|
segment_mapping[mapping_index] = segment_id + 1; |
|
roi_map->delta_q[segment_id] = delta_q; |
|
segment_id++; |
|
} |
|
} |
|
|
|
roi_map->rows = (frame_height + block_size - 1) / block_size; |
|
roi_map->cols = (frame_width + block_size - 1) / block_size; |
|
roi_map->roi_map = av_mallocz_array(roi_map->rows * roi_map->cols, sizeof(*roi_map->roi_map)); |
|
if (!roi_map->roi_map) { |
|
av_log(avctx, AV_LOG_ERROR, "roi_map alloc failed.\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
/* This list must be iterated in reverse, so for the case that |
|
* two regions are overlapping, the more important area takes effect. |
|
*/ |
|
for (int i = nb_rois - 1; i >= 0; i--) { |
|
int delta_q; |
|
int mapping_value; |
|
int starty, endy, startx, endx; |
|
|
|
roi = (const AVRegionOfInterest*)(sd->data + self_size * i); |
|
|
|
starty = av_clip(roi->top / block_size, 0, roi_map->rows); |
|
endy = av_clip((roi->bottom + block_size - 1) / block_size, 0, roi_map->rows); |
|
startx = av_clip(roi->left / block_size, 0, roi_map->cols); |
|
endx = av_clip((roi->right + block_size - 1) / block_size, 0, roi_map->cols); |
|
|
|
delta_q = (int)(roi->qoffset.num * 1.0f / roi->qoffset.den * MAX_DELTA_Q); |
|
delta_q = av_clip(delta_q, -MAX_DELTA_Q, MAX_DELTA_Q); |
|
|
|
mapping_value = segment_mapping[delta_q + MAX_DELTA_Q]; |
|
if (mapping_value) { |
|
for (int y = starty; y < endy; y++) |
|
for (int x = startx; x < endx; x++) |
|
roi_map->roi_map[x + y * roi_map->cols] = mapping_value - 1; |
|
} |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int vp9_encode_set_roi(AVCodecContext *avctx, int frame_width, int frame_height, const AVFrameSideData *sd) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
|
|
#ifdef VPX_CTRL_VP9E_SET_ROI_MAP |
|
int version = vpx_codec_version(); |
|
int major = VPX_VERSION_MAJOR(version); |
|
int minor = VPX_VERSION_MINOR(version); |
|
int patch = VPX_VERSION_PATCH(version); |
|
|
|
if (major > 1 || (major == 1 && minor > 8) || (major == 1 && minor == 8 && patch >= 1)) { |
|
vpx_roi_map_t roi_map; |
|
const int segment_cnt = 8; |
|
const int block_size = 8; |
|
int ret; |
|
|
|
if (ctx->aq_mode > 0 || ctx->cpu_used < 5 || ctx->deadline != VPX_DL_REALTIME) { |
|
if (!ctx->roi_warned) { |
|
ctx->roi_warned = 1; |
|
av_log(avctx, AV_LOG_WARNING, "ROI is only enabled when aq_mode is 0, cpu_used >= 5 " |
|
"and deadline is REALTIME, so skipping ROI.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
} |
|
|
|
ret = set_roi_map(avctx, sd, frame_width, frame_height, &roi_map, block_size, segment_cnt); |
|
if (ret) { |
|
log_encoder_error(avctx, "Failed to set_roi_map.\n"); |
|
return ret; |
|
} |
|
|
|
memset(roi_map.ref_frame, -1, sizeof(roi_map.ref_frame)); |
|
|
|
if (vpx_codec_control(&ctx->encoder, VP9E_SET_ROI_MAP, &roi_map)) { |
|
log_encoder_error(avctx, "Failed to set VP9E_SET_ROI_MAP codec control.\n"); |
|
ret = AVERROR_INVALIDDATA; |
|
} |
|
av_freep(&roi_map.roi_map); |
|
return ret; |
|
} |
|
#endif |
|
|
|
if (!ctx->roi_warned) { |
|
ctx->roi_warned = 1; |
|
av_log(avctx, AV_LOG_WARNING, "ROI is not supported, please upgrade libvpx to version >= 1.8.1. " |
|
"You may need to rebuild ffmpeg.\n"); |
|
} |
|
return 0; |
|
} |
|
|
|
static int vp8_encode_set_roi(AVCodecContext *avctx, int frame_width, int frame_height, const AVFrameSideData *sd) |
|
{ |
|
vpx_roi_map_t roi_map; |
|
const int segment_cnt = 4; |
|
const int block_size = 16; |
|
VPxContext *ctx = avctx->priv_data; |
|
|
|
int ret = set_roi_map(avctx, sd, frame_width, frame_height, &roi_map, block_size, segment_cnt); |
|
if (ret) { |
|
log_encoder_error(avctx, "Failed to set_roi_map.\n"); |
|
return ret; |
|
} |
|
|
|
if (vpx_codec_control(&ctx->encoder, VP8E_SET_ROI_MAP, &roi_map)) { |
|
log_encoder_error(avctx, "Failed to set VP8E_SET_ROI_MAP codec control.\n"); |
|
ret = AVERROR_INVALIDDATA; |
|
} |
|
|
|
av_freep(&roi_map.roi_map); |
|
return ret; |
|
} |
|
|
|
static int realloc_alpha_uv(AVCodecContext *avctx, int width, int height) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
struct vpx_image *rawimg_alpha = &ctx->rawimg_alpha; |
|
unsigned char **planes = rawimg_alpha->planes; |
|
int *stride = rawimg_alpha->stride; |
|
|
|
if (!planes[VPX_PLANE_U] || |
|
!planes[VPX_PLANE_V] || |
|
width != (int)rawimg_alpha->d_w || |
|
height != (int)rawimg_alpha->d_h) { |
|
av_freep(&planes[VPX_PLANE_U]); |
|
av_freep(&planes[VPX_PLANE_V]); |
|
|
|
vpx_img_wrap(rawimg_alpha, VPX_IMG_FMT_I420, width, height, 1, |
|
(unsigned char*)1); |
|
planes[VPX_PLANE_U] = av_malloc_array(stride[VPX_PLANE_U], height); |
|
planes[VPX_PLANE_V] = av_malloc_array(stride[VPX_PLANE_V], height); |
|
if (!planes[VPX_PLANE_U] || !planes[VPX_PLANE_V]) |
|
return AVERROR(ENOMEM); |
|
|
|
memset(planes[VPX_PLANE_U], 0x80, stride[VPX_PLANE_U] * height); |
|
memset(planes[VPX_PLANE_V], 0x80, stride[VPX_PLANE_V] * height); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int vpx_encode(AVCodecContext *avctx, AVPacket *pkt, |
|
const AVFrame *frame, int *got_packet) |
|
{ |
|
VPxContext *ctx = avctx->priv_data; |
|
struct vpx_image *rawimg = NULL; |
|
struct vpx_image *rawimg_alpha = NULL; |
|
int64_t timestamp = 0; |
|
int res, coded_size; |
|
vpx_enc_frame_flags_t flags = 0; |
|
|
|
if (frame) { |
|
const AVFrameSideData *sd = av_frame_get_side_data(frame, AV_FRAME_DATA_REGIONS_OF_INTEREST); |
|
rawimg = &ctx->rawimg; |
|
rawimg->planes[VPX_PLANE_Y] = frame->data[0]; |
|
rawimg->planes[VPX_PLANE_U] = frame->data[1]; |
|
rawimg->planes[VPX_PLANE_V] = frame->data[2]; |
|
rawimg->stride[VPX_PLANE_Y] = frame->linesize[0]; |
|
rawimg->stride[VPX_PLANE_U] = frame->linesize[1]; |
|
rawimg->stride[VPX_PLANE_V] = frame->linesize[2]; |
|
if (ctx->is_alpha) { |
|
rawimg_alpha = &ctx->rawimg_alpha; |
|
res = realloc_alpha_uv(avctx, frame->width, frame->height); |
|
if (res < 0) |
|
return res; |
|
rawimg_alpha->planes[VPX_PLANE_Y] = frame->data[3]; |
|
rawimg_alpha->stride[VPX_PLANE_Y] = frame->linesize[3]; |
|
} |
|
timestamp = frame->pts; |
|
#if VPX_IMAGE_ABI_VERSION >= 4 |
|
switch (frame->color_range) { |
|
case AVCOL_RANGE_MPEG: |
|
rawimg->range = VPX_CR_STUDIO_RANGE; |
|
break; |
|
case AVCOL_RANGE_JPEG: |
|
rawimg->range = VPX_CR_FULL_RANGE; |
|
break; |
|
} |
|
#endif |
|
if (frame->pict_type == AV_PICTURE_TYPE_I) |
|
flags |= VPX_EFLAG_FORCE_KF; |
|
if (CONFIG_LIBVPX_VP8_ENCODER && avctx->codec_id == AV_CODEC_ID_VP8 && frame->metadata) { |
|
AVDictionaryEntry* en = av_dict_get(frame->metadata, "vp8-flags", NULL, 0); |
|
if (en) { |
|
flags |= strtoul(en->value, NULL, 10); |
|
} |
|
} |
|
|
|
if (sd) { |
|
if (avctx->codec_id == AV_CODEC_ID_VP8) { |
|
vp8_encode_set_roi(avctx, frame->width, frame->height, sd); |
|
} else { |
|
vp9_encode_set_roi(avctx, frame->width, frame->height, sd); |
|
} |
|
} |
|
} |
|
|
|
res = vpx_codec_encode(&ctx->encoder, rawimg, timestamp, |
|
avctx->ticks_per_frame, flags, ctx->deadline); |
|
if (res != VPX_CODEC_OK) { |
|
log_encoder_error(avctx, "Error encoding frame"); |
|
return AVERROR_INVALIDDATA; |
|
} |
|
|
|
if (ctx->is_alpha) { |
|
res = vpx_codec_encode(&ctx->encoder_alpha, rawimg_alpha, timestamp, |
|
avctx->ticks_per_frame, flags, ctx->deadline); |
|
if (res != VPX_CODEC_OK) { |
|
log_encoder_error(avctx, "Error encoding alpha frame"); |
|
return AVERROR_INVALIDDATA; |
|
} |
|
} |
|
|
|
coded_size = queue_frames(avctx, pkt); |
|
|
|
if (!frame && avctx->flags & AV_CODEC_FLAG_PASS1) { |
|
unsigned int b64_size = AV_BASE64_SIZE(ctx->twopass_stats.sz); |
|
|
|
avctx->stats_out = av_malloc(b64_size); |
|
if (!avctx->stats_out) { |
|
av_log(avctx, AV_LOG_ERROR, "Stat buffer alloc (%d bytes) failed\n", |
|
b64_size); |
|
return AVERROR(ENOMEM); |
|
} |
|
av_base64_encode(avctx->stats_out, b64_size, ctx->twopass_stats.buf, |
|
ctx->twopass_stats.sz); |
|
} |
|
|
|
*got_packet = !!coded_size; |
|
return 0; |
|
} |
|
|
|
#define OFFSET(x) offsetof(VPxContext, x) |
|
#define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM |
|
|
|
#define COMMON_OPTIONS \ |
|
{ "lag-in-frames", "Number of frames to look ahead for " \ |
|
"alternate reference frame selection", OFFSET(lag_in_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE}, \ |
|
{ "arnr-maxframes", "altref noise reduction max frame count", OFFSET(arnr_max_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE}, \ |
|
{ "arnr-strength", "altref noise reduction filter strength", OFFSET(arnr_strength), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE}, \ |
|
{ "arnr-type", "altref noise reduction filter type", OFFSET(arnr_type), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE, "arnr_type"}, \ |
|
{ "backward", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = 1}, 0, 0, VE, "arnr_type" }, \ |
|
{ "forward", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = 2}, 0, 0, VE, "arnr_type" }, \ |
|
{ "centered", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = 3}, 0, 0, VE, "arnr_type" }, \ |
|
{ "tune", "Tune the encoding to a specific scenario", OFFSET(tune), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE, "tune"}, \ |
|
{ "psnr", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = VP8_TUNE_PSNR}, 0, 0, VE, "tune"}, \ |
|
{ "ssim", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = VP8_TUNE_SSIM}, 0, 0, VE, "tune"}, \ |
|
{ "deadline", "Time to spend encoding, in microseconds.", OFFSET(deadline), AV_OPT_TYPE_INT, {.i64 = VPX_DL_GOOD_QUALITY}, INT_MIN, INT_MAX, VE, "quality"}, \ |
|
{ "best", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = VPX_DL_BEST_QUALITY}, 0, 0, VE, "quality"}, \ |
|
{ "good", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = VPX_DL_GOOD_QUALITY}, 0, 0, VE, "quality"}, \ |
|
{ "realtime", NULL, 0, AV_OPT_TYPE_CONST, {.i64 = VPX_DL_REALTIME}, 0, 0, VE, "quality"}, \ |
|
{ "error-resilient", "Error resilience configuration", OFFSET(error_resilient), AV_OPT_TYPE_FLAGS, {.i64 = 0}, INT_MIN, INT_MAX, VE, "er"}, \ |
|
{ "max-intra-rate", "Maximum I-frame bitrate (pct) 0=unlimited", OFFSET(max_intra_rate), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE}, \ |
|
{ "default", "Improve resiliency against losses of whole frames", 0, AV_OPT_TYPE_CONST, {.i64 = VPX_ERROR_RESILIENT_DEFAULT}, 0, 0, VE, "er"}, \ |
|
{ "partitions", "The frame partitions are independently decodable " \ |
|
"by the bool decoder, meaning that partitions can be decoded even " \ |
|
"though earlier partitions have been lost. Note that intra prediction" \ |
|
" is still done over the partition boundary.", 0, AV_OPT_TYPE_CONST, {.i64 = VPX_ERROR_RESILIENT_PARTITIONS}, 0, 0, VE, "er"}, \ |
|
{ "crf", "Select the quality for constant quality mode", offsetof(VPxContext, crf), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 63, VE }, \ |
|
{ "static-thresh", "A change threshold on blocks below which they will be skipped by the encoder", OFFSET(static_thresh), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, VE }, \ |
|
{ "drop-threshold", "Frame drop threshold", offsetof(VPxContext, drop_threshold), AV_OPT_TYPE_INT, {.i64 = 0 }, INT_MIN, INT_MAX, VE }, \ |
|
{ "noise-sensitivity", "Noise sensitivity", OFFSET(noise_sensitivity), AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 4, VE}, \ |
|
{ "undershoot-pct", "Datarate undershoot (min) target (%)", OFFSET(rc_undershoot_pct), AV_OPT_TYPE_INT, { .i64 = -1 }, -1, 100, VE }, \ |
|
{ "overshoot-pct", "Datarate overshoot (max) target (%)", OFFSET(rc_overshoot_pct), AV_OPT_TYPE_INT, { .i64 = -1 }, -1, 1000, VE }, \ |
|
|
|
#define LEGACY_OPTIONS \ |
|
{"speed", "", offsetof(VPxContext, cpu_used), AV_OPT_TYPE_INT, {.i64 = 1}, -16, 16, VE}, \ |
|
{"quality", "", offsetof(VPxContext, deadline), AV_OPT_TYPE_INT, {.i64 = VPX_DL_GOOD_QUALITY}, INT_MIN, INT_MAX, VE, "quality"}, \ |
|
{"vp8flags", "", offsetof(VPxContext, flags), AV_OPT_TYPE_FLAGS, {.i64 = 0}, 0, UINT_MAX, VE, "flags"}, \ |
|
{"error_resilient", "enable error resilience", 0, AV_OPT_TYPE_CONST, {.i64 = VP8F_ERROR_RESILIENT}, INT_MIN, INT_MAX, VE, "flags"}, \ |
|
{"altref", "enable use of alternate reference frames (VP8/2-pass only)", 0, AV_OPT_TYPE_CONST, {.i64 = VP8F_AUTO_ALT_REF}, INT_MIN, INT_MAX, VE, "flags"}, \ |
|
{"arnr_max_frames", "altref noise reduction max frame count", offsetof(VPxContext, arnr_max_frames), AV_OPT_TYPE_INT, {.i64 = 0}, 0, 15, VE}, \ |
|
{"arnr_strength", "altref noise reduction filter strength", offsetof(VPxContext, arnr_strength), AV_OPT_TYPE_INT, {.i64 = 3}, 0, 6, VE}, \ |
|
{"arnr_type", "altref noise reduction filter type", offsetof(VPxContext, arnr_type), AV_OPT_TYPE_INT, {.i64 = 3}, 1, 3, VE}, \ |
|
{"rc_lookahead", "Number of frames to look ahead for alternate reference frame selection", offsetof(VPxContext, lag_in_frames), AV_OPT_TYPE_INT, {.i64 = 25}, 0, 25, VE}, \ |
|
{"sharpness", "Increase sharpness at the expense of lower PSNR", offsetof(VPxContext, sharpness), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 7, VE}, |
|
|
|
#if CONFIG_LIBVPX_VP8_ENCODER |
|
static const AVOption vp8_options[] = { |
|
COMMON_OPTIONS |
|
{ "auto-alt-ref", "Enable use of alternate reference " |
|
"frames (2-pass only)", OFFSET(auto_alt_ref), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 2, VE}, |
|
{ "cpu-used", "Quality/Speed ratio modifier", OFFSET(cpu_used), AV_OPT_TYPE_INT, {.i64 = 1}, -16, 16, VE}, |
|
{ "ts-parameters", "Temporal scaling configuration using a " |
|
":-separated list of key=value parameters", OFFSET(vp8_ts_parameters), AV_OPT_TYPE_DICT, {.str=NULL}, 0, 0, VE}, |
|
LEGACY_OPTIONS |
|
{ NULL } |
|
}; |
|
#endif |
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
static const AVOption vp9_options[] = { |
|
COMMON_OPTIONS |
|
{ "auto-alt-ref", "Enable use of alternate reference " |
|
"frames (2-pass only)", OFFSET(auto_alt_ref), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE}, |
|
{ "cpu-used", "Quality/Speed ratio modifier", OFFSET(cpu_used), AV_OPT_TYPE_INT, {.i64 = 1}, -8, 8, VE}, |
|
{ "lossless", "Lossless mode", OFFSET(lossless), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 1, VE}, |
|
{ "tile-columns", "Number of tile columns to use, log2", OFFSET(tile_columns), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE}, |
|
{ "tile-rows", "Number of tile rows to use, log2", OFFSET(tile_rows), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 2, VE}, |
|
{ "frame-parallel", "Enable frame parallel decodability features", OFFSET(frame_parallel), AV_OPT_TYPE_BOOL,{.i64 = -1}, -1, 1, VE}, |
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
{ "aq-mode", "adaptive quantization mode", OFFSET(aq_mode), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 4, VE, "aq_mode"}, |
|
#else |
|
{ "aq-mode", "adaptive quantization mode", OFFSET(aq_mode), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 3, VE, "aq_mode"}, |
|
#endif |
|
{ "none", "Aq not used", 0, AV_OPT_TYPE_CONST, {.i64 = 0}, 0, 0, VE, "aq_mode" }, |
|
{ "variance", "Variance based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 1}, 0, 0, VE, "aq_mode" }, |
|
{ "complexity", "Complexity based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 2}, 0, 0, VE, "aq_mode" }, |
|
{ "cyclic", "Cyclic Refresh Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 3}, 0, 0, VE, "aq_mode" }, |
|
#if VPX_ENCODER_ABI_VERSION >= 12 |
|
{ "equator360", "360 video Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 4}, 0, 0, VE, "aq_mode" }, |
|
{"level", "Specify level", OFFSET(level), AV_OPT_TYPE_FLOAT, {.dbl=-1}, -1, 6.2, VE}, |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_ROW_MT |
|
{"row-mt", "Row based multi-threading", OFFSET(row_mt), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE}, |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TUNE_CONTENT |
|
#if VPX_ENCODER_ABI_VERSION >= 14 |
|
{ "tune-content", "Tune content type", OFFSET(tune_content), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 2, VE, "tune_content" }, |
|
#else |
|
{ "tune-content", "Tune content type", OFFSET(tune_content), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 1, VE, "tune_content" }, |
|
#endif |
|
{ "default", "Regular video content", 0, AV_OPT_TYPE_CONST, {.i64 = 0}, 0, 0, VE, "tune_content" }, |
|
{ "screen", "Screen capture content", 0, AV_OPT_TYPE_CONST, {.i64 = 1}, 0, 0, VE, "tune_content" }, |
|
#if VPX_ENCODER_ABI_VERSION >= 14 |
|
{ "film", "Film content; improves grain retention", 0, AV_OPT_TYPE_CONST, {.i64 = 2}, 0, 0, VE, "tune_content" }, |
|
#endif |
|
#endif |
|
#if VPX_ENCODER_ABI_VERSION >= 14 |
|
{ "corpus-complexity", "corpus vbr complexity midpoint", OFFSET(corpus_complexity), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 10000, VE }, |
|
#endif |
|
#ifdef VPX_CTRL_VP9E_SET_TPL |
|
{ "enable-tpl", "Enable temporal dependency model", OFFSET(tpl_model), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE }, |
|
#endif |
|
LEGACY_OPTIONS |
|
{ NULL } |
|
}; |
|
#endif |
|
|
|
#undef COMMON_OPTIONS |
|
#undef LEGACY_OPTIONS |
|
|
|
static const AVCodecDefault defaults[] = { |
|
{ "b", "0" }, |
|
{ "qmin", "-1" }, |
|
{ "qmax", "-1" }, |
|
{ "g", "-1" }, |
|
{ "keyint_min", "-1" }, |
|
{ NULL }, |
|
}; |
|
|
|
#if CONFIG_LIBVPX_VP8_ENCODER |
|
static av_cold int vp8_init(AVCodecContext *avctx) |
|
{ |
|
return vpx_init(avctx, vpx_codec_vp8_cx()); |
|
} |
|
|
|
static const AVClass class_vp8 = { |
|
.class_name = "libvpx-vp8 encoder", |
|
.item_name = av_default_item_name, |
|
.option = vp8_options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
}; |
|
|
|
AVCodec ff_libvpx_vp8_encoder = { |
|
.name = "libvpx", |
|
.long_name = NULL_IF_CONFIG_SMALL("libvpx VP8"), |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_VP8, |
|
.priv_data_size = sizeof(VPxContext), |
|
.init = vp8_init, |
|
.encode2 = vpx_encode, |
|
.close = vpx_free, |
|
.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_AUTO_THREADS, |
|
.pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_NONE }, |
|
.priv_class = &class_vp8, |
|
.defaults = defaults, |
|
.wrapper_name = "libvpx", |
|
}; |
|
#endif /* CONFIG_LIBVPX_VP8_ENCODER */ |
|
|
|
#if CONFIG_LIBVPX_VP9_ENCODER |
|
static av_cold int vp9_init(AVCodecContext *avctx) |
|
{ |
|
return vpx_init(avctx, vpx_codec_vp9_cx()); |
|
} |
|
|
|
static const AVClass class_vp9 = { |
|
.class_name = "libvpx-vp9 encoder", |
|
.item_name = av_default_item_name, |
|
.option = vp9_options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
}; |
|
|
|
AVCodec ff_libvpx_vp9_encoder = { |
|
.name = "libvpx-vp9", |
|
.long_name = NULL_IF_CONFIG_SMALL("libvpx VP9"), |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_VP9, |
|
.priv_data_size = sizeof(VPxContext), |
|
.init = vp9_init, |
|
.encode2 = vpx_encode, |
|
.close = vpx_free, |
|
.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_AUTO_THREADS, |
|
.profiles = NULL_IF_CONFIG_SMALL(ff_vp9_profiles), |
|
.priv_class = &class_vp9, |
|
.defaults = defaults, |
|
.init_static_data = ff_vp9_init_static, |
|
.wrapper_name = "libvpx", |
|
}; |
|
#endif /* CONFIG_LIBVPX_VP9_ENCODER */
|
|
|