You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1132 lines
36 KiB

/*
* AV1 video decoder
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/film_grain_params.h"
#include "libavutil/pixdesc.h"
#include "avcodec.h"
#include "av1dec.h"
#include "bytestream.h"
#include "hwconfig.h"
#include "internal.h"
#include "profiles.h"
static uint32_t inverse_recenter(int r, uint32_t v)
{
if (v > 2 * r)
return v;
else if (v & 1)
return r - ((v + 1) >> 1);
else
return r + (v >> 1);
}
static uint32_t decode_unsigned_subexp_with_ref(uint32_t sub_exp,
int mx, int r)
{
if ((r << 1) <= mx) {
return inverse_recenter(r, sub_exp);
} else {
return mx - 1 - inverse_recenter(mx - 1 - r, sub_exp);
}
}
static int32_t decode_signed_subexp_with_ref(uint32_t sub_exp, int low,
int high, int r)
{
int32_t x = decode_unsigned_subexp_with_ref(sub_exp, high - low, r - low);
return x + low;
}
static void read_global_param(AV1DecContext *s, int type, int ref, int idx)
{
uint8_t primary_frame, prev_frame;
uint32_t abs_bits, prec_bits, round, prec_diff, sub, mx;
int32_t r, prev_gm_param;
primary_frame = s->raw_frame_header->primary_ref_frame;
prev_frame = s->raw_frame_header->ref_frame_idx[primary_frame];
abs_bits = AV1_GM_ABS_ALPHA_BITS;
prec_bits = AV1_GM_ALPHA_PREC_BITS;
/* setup_past_independence() sets PrevGmParams to default values. We can
* simply point to the current's frame gm_params as they will be initialized
* with defaults at this point.
*/
if (s->raw_frame_header->primary_ref_frame == AV1_PRIMARY_REF_NONE)
prev_gm_param = s->cur_frame.gm_params[ref][idx];
else
prev_gm_param = s->ref[prev_frame].gm_params[ref][idx];
if (idx < 2) {
if (type == AV1_WARP_MODEL_TRANSLATION) {
abs_bits = AV1_GM_ABS_TRANS_ONLY_BITS -
!s->raw_frame_header->allow_high_precision_mv;
prec_bits = AV1_GM_TRANS_ONLY_PREC_BITS -
!s->raw_frame_header->allow_high_precision_mv;
} else {
abs_bits = AV1_GM_ABS_TRANS_BITS;
prec_bits = AV1_GM_TRANS_PREC_BITS;
}
}
round = (idx % 3) == 2 ? (1 << AV1_WARPEDMODEL_PREC_BITS) : 0;
prec_diff = AV1_WARPEDMODEL_PREC_BITS - prec_bits;
sub = (idx % 3) == 2 ? (1 << prec_bits) : 0;
mx = 1 << abs_bits;
r = (prev_gm_param >> prec_diff) - sub;
s->cur_frame.gm_params[ref][idx] =
(decode_signed_subexp_with_ref(s->raw_frame_header->gm_params[ref][idx],
-mx, mx + 1, r) << prec_diff) + round;
}
/**
* update gm type/params, since cbs already implemented part of this funcation,
* so we don't need to full implement spec.
*/
static void global_motion_params(AV1DecContext *s)
{
const AV1RawFrameHeader *header = s->raw_frame_header;
int type, ref;
for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
s->cur_frame.gm_type[ref] = AV1_WARP_MODEL_IDENTITY;
for (int i = 0; i < 6; i++)
s->cur_frame.gm_params[ref][i] = (i % 3 == 2) ?
1 << AV1_WARPEDMODEL_PREC_BITS : 0;
}
if (header->frame_type == AV1_FRAME_KEY ||
header->frame_type == AV1_FRAME_INTRA_ONLY)
return;
for (ref = AV1_REF_FRAME_LAST; ref <= AV1_REF_FRAME_ALTREF; ref++) {
if (header->is_global[ref]) {
if (header->is_rot_zoom[ref]) {
type = AV1_WARP_MODEL_ROTZOOM;
} else {
type = header->is_translation[ref] ? AV1_WARP_MODEL_TRANSLATION
: AV1_WARP_MODEL_AFFINE;
}
} else {
type = AV1_WARP_MODEL_IDENTITY;
}
s->cur_frame.gm_type[ref] = type;
if (type >= AV1_WARP_MODEL_ROTZOOM) {
read_global_param(s, type, ref, 2);
read_global_param(s, type, ref, 3);
if (type == AV1_WARP_MODEL_AFFINE) {
read_global_param(s, type, ref, 4);
read_global_param(s, type, ref, 5);
} else {
s->cur_frame.gm_params[ref][4] = -s->cur_frame.gm_params[ref][3];
s->cur_frame.gm_params[ref][5] = s->cur_frame.gm_params[ref][2];
}
}
if (type >= AV1_WARP_MODEL_TRANSLATION) {
read_global_param(s, type, ref, 0);
read_global_param(s, type, ref, 1);
}
}
}
static int get_relative_dist(const AV1RawSequenceHeader *seq,
unsigned int a, unsigned int b)
{
unsigned int diff = a - b;
unsigned int m = 1 << seq->order_hint_bits_minus_1;
return (diff & (m - 1)) - (diff & m);
}
static void skip_mode_params(AV1DecContext *s)
{
const AV1RawFrameHeader *header = s->raw_frame_header;
const AV1RawSequenceHeader *seq = s->raw_seq;
int forward_idx, backward_idx;
int forward_hint, backward_hint;
int second_forward_idx, second_forward_hint;
int ref_hint, dist, i;
if (!header->skip_mode_present)
return;
forward_idx = -1;
backward_idx = -1;
for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
ref_hint = s->ref[header->ref_frame_idx[i]].raw_frame_header->order_hint;
dist = get_relative_dist(seq, ref_hint, header->order_hint);
if (dist < 0) {
if (forward_idx < 0 ||
get_relative_dist(seq, ref_hint, forward_hint) > 0) {
forward_idx = i;
forward_hint = ref_hint;
}
} else if (dist > 0) {
if (backward_idx < 0 ||
get_relative_dist(seq, ref_hint, backward_hint) < 0) {
backward_idx = i;
backward_hint = ref_hint;
}
}
}
if (forward_idx < 0) {
return;
} else if (backward_idx >= 0) {
s->cur_frame.skip_mode_frame_idx[0] =
AV1_REF_FRAME_LAST + FFMIN(forward_idx, backward_idx);
s->cur_frame.skip_mode_frame_idx[1] =
AV1_REF_FRAME_LAST + FFMAX(forward_idx, backward_idx);
return;
}
second_forward_idx = -1;
for (i = 0; i < AV1_REFS_PER_FRAME; i++) {
ref_hint = s->ref[header->ref_frame_idx[i]].raw_frame_header->order_hint;
if (get_relative_dist(seq, ref_hint, forward_hint) < 0) {
if (second_forward_idx < 0 ||
get_relative_dist(seq, ref_hint, second_forward_hint) > 0) {
second_forward_idx = i;
second_forward_hint = ref_hint;
}
}
}
if (second_forward_idx < 0)
return;
s->cur_frame.skip_mode_frame_idx[0] =
AV1_REF_FRAME_LAST + FFMIN(forward_idx, second_forward_idx);
s->cur_frame.skip_mode_frame_idx[1] =
AV1_REF_FRAME_LAST + FFMAX(forward_idx, second_forward_idx);
}
static void coded_lossless_param(AV1DecContext *s)
{
const AV1RawFrameHeader *header = s->raw_frame_header;
int i;
if (header->delta_q_y_dc || header->delta_q_u_ac ||
header->delta_q_u_dc || header->delta_q_v_ac ||
header->delta_q_v_dc) {
s->cur_frame.coded_lossless = 0;
return;
}
s->cur_frame.coded_lossless = 1;
for (i = 0; i < AV1_MAX_SEGMENTS; i++) {
int qindex;
if (header->feature_enabled[i][AV1_SEG_LVL_ALT_Q]) {
qindex = (header->base_q_idx +
header->feature_value[i][AV1_SEG_LVL_ALT_Q]);
} else {
qindex = header->base_q_idx;
}
qindex = av_clip_uintp2(qindex, 8);
if (qindex) {
s->cur_frame.coded_lossless = 0;
return;
}
}
}
static void load_grain_params(AV1DecContext *s)
{
const AV1RawFrameHeader *header = s->raw_frame_header;
const AV1RawFilmGrainParams *film_grain = &header->film_grain, *src;
AV1RawFilmGrainParams *dst = &s->cur_frame.film_grain;
if (!film_grain->apply_grain)
return;
if (film_grain->update_grain) {
memcpy(dst, film_grain, sizeof(*dst));
return;
}
src = &s->ref[film_grain->film_grain_params_ref_idx].film_grain;
memcpy(dst, src, sizeof(*dst));
dst->grain_seed = film_grain->grain_seed;
}
static int init_tile_data(AV1DecContext *s)
{
int cur_tile_num =
s->raw_frame_header->tile_cols * s->raw_frame_header->tile_rows;
if (s->tile_num < cur_tile_num) {
int ret = av_reallocp_array(&s->tile_group_info, cur_tile_num,
sizeof(TileGroupInfo));
if (ret < 0) {
s->tile_num = 0;
return ret;
}
}
s->tile_num = cur_tile_num;
return 0;
}
static int get_tiles_info(AVCodecContext *avctx, const AV1RawTileGroup *tile_group)
{
AV1DecContext *s = avctx->priv_data;
GetByteContext gb;
uint16_t tile_num, tile_row, tile_col;
uint32_t size = 0, size_bytes = 0;
bytestream2_init(&gb, tile_group->tile_data.data,
tile_group->tile_data.data_size);
s->tg_start = tile_group->tg_start;
s->tg_end = tile_group->tg_end;
for (tile_num = tile_group->tg_start; tile_num <= tile_group->tg_end; tile_num++) {
tile_row = tile_num / s->raw_frame_header->tile_cols;
tile_col = tile_num % s->raw_frame_header->tile_cols;
if (tile_num == tile_group->tg_end) {
s->tile_group_info[tile_num].tile_size = bytestream2_get_bytes_left(&gb);
s->tile_group_info[tile_num].tile_offset = bytestream2_tell(&gb);
s->tile_group_info[tile_num].tile_row = tile_row;
s->tile_group_info[tile_num].tile_column = tile_col;
return 0;
}
size_bytes = s->raw_frame_header->tile_size_bytes_minus1 + 1;
if (bytestream2_get_bytes_left(&gb) < size_bytes)
return AVERROR_INVALIDDATA;
size = 0;
for (int i = 0; i < size_bytes; i++)
size |= bytestream2_get_byteu(&gb) << 8 * i;
if (bytestream2_get_bytes_left(&gb) <= size)
return AVERROR_INVALIDDATA;
size++;
s->tile_group_info[tile_num].tile_size = size;
s->tile_group_info[tile_num].tile_offset = bytestream2_tell(&gb);
s->tile_group_info[tile_num].tile_row = tile_row;
s->tile_group_info[tile_num].tile_column = tile_col;
bytestream2_skipu(&gb, size);
}
return 0;
}
static int get_pixel_format(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
const AV1RawSequenceHeader *seq = s->raw_seq;
uint8_t bit_depth;
int ret;
enum AVPixelFormat pix_fmt = AV_PIX_FMT_NONE;
#define HWACCEL_MAX (CONFIG_AV1_DXVA2_HWACCEL + \
CONFIG_AV1_D3D11VA_HWACCEL * 2 + \
CONFIG_AV1_NVDEC_HWACCEL + \
CONFIG_AV1_VAAPI_HWACCEL)
enum AVPixelFormat pix_fmts[HWACCEL_MAX + 2], *fmtp = pix_fmts;
if (seq->seq_profile == 2 && seq->color_config.high_bitdepth)
bit_depth = seq->color_config.twelve_bit ? 12 : 10;
else if (seq->seq_profile <= 2)
bit_depth = seq->color_config.high_bitdepth ? 10 : 8;
else {
av_log(avctx, AV_LOG_ERROR,
"Unknown AV1 profile %d.\n", seq->seq_profile);
return -1;
}
if (!seq->color_config.mono_chrome) {
// 4:4:4 x:0 y:0, 4:2:2 x:1 y:0, 4:2:0 x:1 y:1
if (seq->color_config.subsampling_x == 0 &&
seq->color_config.subsampling_y == 0) {
if (bit_depth == 8)
pix_fmt = AV_PIX_FMT_YUV444P;
else if (bit_depth == 10)
pix_fmt = AV_PIX_FMT_YUV444P10;
else if (bit_depth == 12)
pix_fmt = AV_PIX_FMT_YUV444P12;
else
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
} else if (seq->color_config.subsampling_x == 1 &&
seq->color_config.subsampling_y == 0) {
if (bit_depth == 8)
pix_fmt = AV_PIX_FMT_YUV422P;
else if (bit_depth == 10)
pix_fmt = AV_PIX_FMT_YUV422P10;
else if (bit_depth == 12)
pix_fmt = AV_PIX_FMT_YUV422P12;
else
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
} else if (seq->color_config.subsampling_x == 1 &&
seq->color_config.subsampling_y == 1) {
if (bit_depth == 8)
pix_fmt = AV_PIX_FMT_YUV420P;
else if (bit_depth == 10)
pix_fmt = AV_PIX_FMT_YUV420P10;
else if (bit_depth == 12)
pix_fmt = AV_PIX_FMT_YUV420P12;
else
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
}
} else {
if (bit_depth == 8)
pix_fmt = AV_PIX_FMT_GRAY8;
else if (bit_depth == 10)
pix_fmt = AV_PIX_FMT_GRAY10;
else if (bit_depth == 12)
pix_fmt = AV_PIX_FMT_GRAY12;
else
av_log(avctx, AV_LOG_WARNING, "Unknown AV1 pixel format.\n");
}
av_log(avctx, AV_LOG_DEBUG, "AV1 decode get format: %s.\n",
av_get_pix_fmt_name(pix_fmt));
if (pix_fmt == AV_PIX_FMT_NONE)
return -1;
s->pix_fmt = pix_fmt;
switch (s->pix_fmt) {
case AV_PIX_FMT_YUV420P:
#if CONFIG_AV1_DXVA2_HWACCEL
*fmtp++ = AV_PIX_FMT_DXVA2_VLD;
#endif
#if CONFIG_AV1_D3D11VA_HWACCEL
*fmtp++ = AV_PIX_FMT_D3D11VA_VLD;
*fmtp++ = AV_PIX_FMT_D3D11;
#endif
#if CONFIG_AV1_NVDEC_HWACCEL
*fmtp++ = AV_PIX_FMT_CUDA;
#endif
#if CONFIG_AV1_VAAPI_HWACCEL
*fmtp++ = AV_PIX_FMT_VAAPI;
#endif
break;
case AV_PIX_FMT_YUV420P10:
#if CONFIG_AV1_DXVA2_HWACCEL
*fmtp++ = AV_PIX_FMT_DXVA2_VLD;
#endif
#if CONFIG_AV1_D3D11VA_HWACCEL
*fmtp++ = AV_PIX_FMT_D3D11VA_VLD;
*fmtp++ = AV_PIX_FMT_D3D11;
#endif
#if CONFIG_AV1_NVDEC_HWACCEL
*fmtp++ = AV_PIX_FMT_CUDA;
#endif
#if CONFIG_AV1_VAAPI_HWACCEL
*fmtp++ = AV_PIX_FMT_VAAPI;
#endif
break;
case AV_PIX_FMT_GRAY8:
#if CONFIG_AV1_NVDEC_HWACCEL
*fmtp++ = AV_PIX_FMT_CUDA;
#endif
break;
case AV_PIX_FMT_GRAY10:
#if CONFIG_AV1_NVDEC_HWACCEL
*fmtp++ = AV_PIX_FMT_CUDA;
#endif
break;
}
*fmtp++ = s->pix_fmt;
*fmtp = AV_PIX_FMT_NONE;
ret = ff_thread_get_format(avctx, pix_fmts);
if (ret < 0)
return ret;
/**
* check if the HW accel is inited correctly. If not, return un-implemented.
* Since now the av1 decoder doesn't support native decode, if it will be
* implemented in the future, need remove this check.
*/
if (!avctx->hwaccel) {
av_log(avctx, AV_LOG_ERROR, "Your platform doesn't suppport"
" hardware accelerated AV1 decoding.\n");
return AVERROR(ENOSYS);
}
avctx->pix_fmt = ret;
return 0;
}
static void av1_frame_unref(AVCodecContext *avctx, AV1Frame *f)
{
ff_thread_release_buffer(avctx, &f->tf);
av_buffer_unref(&f->hwaccel_priv_buf);
f->hwaccel_picture_private = NULL;
av_buffer_unref(&f->header_ref);
f->raw_frame_header = NULL;
f->spatial_id = f->temporal_id = 0;
memset(f->skip_mode_frame_idx, 0,
2 * sizeof(uint8_t));
memset(&f->film_grain, 0, sizeof(f->film_grain));
f->coded_lossless = 0;
}
static int av1_frame_ref(AVCodecContext *avctx, AV1Frame *dst, const AV1Frame *src)
{
int ret;
ret = ff_thread_ref_frame(&dst->tf, &src->tf);
if (ret < 0)
return ret;
dst->header_ref = av_buffer_ref(src->header_ref);
if (!dst->header_ref)
goto fail;
dst->raw_frame_header = src->raw_frame_header;
if (src->hwaccel_picture_private) {
dst->hwaccel_priv_buf = av_buffer_ref(src->hwaccel_priv_buf);
if (!dst->hwaccel_priv_buf)
goto fail;
dst->hwaccel_picture_private = dst->hwaccel_priv_buf->data;
}
dst->spatial_id = src->spatial_id;
dst->temporal_id = src->temporal_id;
memcpy(dst->gm_type,
src->gm_type,
AV1_NUM_REF_FRAMES * sizeof(uint8_t));
memcpy(dst->gm_params,
src->gm_params,
AV1_NUM_REF_FRAMES * 6 * sizeof(int32_t));
memcpy(dst->skip_mode_frame_idx,
src->skip_mode_frame_idx,
2 * sizeof(uint8_t));
memcpy(&dst->film_grain,
&src->film_grain,
sizeof(dst->film_grain));
dst->coded_lossless = src->coded_lossless;
return 0;
fail:
av1_frame_unref(avctx, dst);
return AVERROR(ENOMEM);
}
static av_cold int av1_decode_free(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
av1_frame_unref(avctx, &s->ref[i]);
av_frame_free(&s->ref[i].tf.f);
}
av1_frame_unref(avctx, &s->cur_frame);
av_frame_free(&s->cur_frame.tf.f);
av_buffer_unref(&s->seq_ref);
av_buffer_unref(&s->header_ref);
av_freep(&s->tile_group_info);
ff_cbs_fragment_free(&s->current_obu);
ff_cbs_close(&s->cbc);
return 0;
}
static int set_context_with_sequence(AVCodecContext *avctx,
const AV1RawSequenceHeader *seq)
{
int width = seq->max_frame_width_minus_1 + 1;
int height = seq->max_frame_height_minus_1 + 1;
avctx->profile = seq->seq_profile;
avctx->level = seq->seq_level_idx[0];
avctx->color_range =
seq->color_config.color_range ? AVCOL_RANGE_JPEG : AVCOL_RANGE_MPEG;
avctx->color_primaries = seq->color_config.color_primaries;
avctx->colorspace = seq->color_config.color_primaries;
avctx->color_trc = seq->color_config.transfer_characteristics;
switch (seq->color_config.chroma_sample_position) {
case AV1_CSP_VERTICAL:
avctx->chroma_sample_location = AVCHROMA_LOC_LEFT;
break;
case AV1_CSP_COLOCATED:
avctx->chroma_sample_location = AVCHROMA_LOC_TOPLEFT;
break;
}
if (avctx->width != width || avctx->height != height) {
int ret = ff_set_dimensions(avctx, width, height);
if (ret < 0)
return ret;
}
avctx->sample_aspect_ratio = (AVRational) { 1, 1 };
if (seq->timing_info.num_units_in_display_tick &&
seq->timing_info.time_scale) {
av_reduce(&avctx->framerate.den, &avctx->framerate.num,
seq->timing_info.num_units_in_display_tick,
seq->timing_info.time_scale,
INT_MAX);
if (seq->timing_info.equal_picture_interval)
avctx->ticks_per_frame = seq->timing_info.num_ticks_per_picture_minus_1 + 1;
}
return 0;
}
static int update_context_with_frame_header(AVCodecContext *avctx,
const AV1RawFrameHeader *header)
{
AVRational aspect_ratio;
int width = header->frame_width_minus_1 + 1;
int height = header->frame_height_minus_1 + 1;
int r_width = header->render_width_minus_1 + 1;
int r_height = header->render_height_minus_1 + 1;
int ret;
if (avctx->width != width || avctx->height != height) {
ret = ff_set_dimensions(avctx, width, height);
if (ret < 0)
return ret;
}
av_reduce(&aspect_ratio.num, &aspect_ratio.den,
(int64_t)height * r_width,
(int64_t)width * r_height,
INT_MAX);
if (av_cmp_q(avctx->sample_aspect_ratio, aspect_ratio)) {
ret = ff_set_sar(avctx, aspect_ratio);
if (ret < 0)
return ret;
}
return 0;
}
static av_cold int av1_decode_init(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
AV1RawSequenceHeader *seq;
int ret;
s->avctx = avctx;
s->pix_fmt = AV_PIX_FMT_NONE;
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++) {
s->ref[i].tf.f = av_frame_alloc();
if (!s->ref[i].tf.f) {
av_log(avctx, AV_LOG_ERROR,
"Failed to allocate reference frame buffer %d.\n", i);
return AVERROR(ENOMEM);
}
}
s->cur_frame.tf.f = av_frame_alloc();
if (!s->cur_frame.tf.f) {
av_log(avctx, AV_LOG_ERROR,
"Failed to allocate current frame buffer.\n");
return AVERROR(ENOMEM);
}
ret = ff_cbs_init(&s->cbc, AV_CODEC_ID_AV1, avctx);
if (ret < 0)
return ret;
if (avctx->extradata && avctx->extradata_size) {
ret = ff_cbs_read_extradata_from_codec(s->cbc,
&s->current_obu,
avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_WARNING, "Failed to read extradata.\n");
return ret;
}
seq = ((CodedBitstreamAV1Context *)(s->cbc->priv_data))->sequence_header;
if (!seq) {
av_log(avctx, AV_LOG_WARNING, "No sequence header available.\n");
goto end;
}
ret = set_context_with_sequence(avctx, seq);
if (ret < 0) {
av_log(avctx, AV_LOG_WARNING, "Failed to set decoder context.\n");
goto end;
}
end:
ff_cbs_fragment_reset(&s->current_obu);
}
return ret;
}
static int av1_frame_alloc(AVCodecContext *avctx, AV1Frame *f)
{
AV1DecContext *s = avctx->priv_data;
AV1RawFrameHeader *header= s->raw_frame_header;
AVFrame *frame;
int ret;
ret = update_context_with_frame_header(avctx, header);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to update context with frame header\n");
return ret;
}
f->header_ref = av_buffer_ref(s->header_ref);
if (!f->header_ref)
return AVERROR(ENOMEM);
f->raw_frame_header = s->raw_frame_header;
if ((ret = ff_thread_get_buffer(avctx, &f->tf, AV_GET_BUFFER_FLAG_REF)) < 0)
goto fail;
frame = f->tf.f;
frame->key_frame = header->frame_type == AV1_FRAME_KEY;
switch (header->frame_type) {
case AV1_FRAME_KEY:
case AV1_FRAME_INTRA_ONLY:
frame->pict_type = AV_PICTURE_TYPE_I;
break;
case AV1_FRAME_INTER:
frame->pict_type = AV_PICTURE_TYPE_P;
break;
case AV1_FRAME_SWITCH:
frame->pict_type = AV_PICTURE_TYPE_SP;
break;
}
if (avctx->hwaccel) {
const AVHWAccel *hwaccel = avctx->hwaccel;
if (hwaccel->frame_priv_data_size) {
f->hwaccel_priv_buf =
av_buffer_allocz(hwaccel->frame_priv_data_size);
if (!f->hwaccel_priv_buf) {
ret = AVERROR(ENOMEM);
goto fail;
}
f->hwaccel_picture_private = f->hwaccel_priv_buf->data;
}
}
return 0;
fail:
av1_frame_unref(avctx, f);
return ret;
}
static int export_film_grain(AVCodecContext *avctx, AVFrame *frame)
{
AV1DecContext *s = avctx->priv_data;
const AV1RawFilmGrainParams *film_grain = &s->cur_frame.film_grain;
AVFilmGrainParams *fgp;
AVFilmGrainAOMParams *aom;
if (!film_grain->apply_grain)
return 0;
fgp = av_film_grain_params_create_side_data(frame);
if (!fgp)
return AVERROR(ENOMEM);
fgp->type = AV_FILM_GRAIN_PARAMS_AV1;
fgp->seed = film_grain->grain_seed;
aom = &fgp->codec.aom;
aom->chroma_scaling_from_luma = film_grain->chroma_scaling_from_luma;
aom->scaling_shift = film_grain->grain_scaling_minus_8 + 8;
aom->ar_coeff_lag = film_grain->ar_coeff_lag;
aom->ar_coeff_shift = film_grain->ar_coeff_shift_minus_6 + 6;
aom->grain_scale_shift = film_grain->grain_scale_shift;
aom->overlap_flag = film_grain->overlap_flag;
aom->limit_output_range = film_grain->clip_to_restricted_range;
aom->num_y_points = film_grain->num_y_points;
for (int i = 0; i < film_grain->num_y_points; i++) {
aom->y_points[i][0] = film_grain->point_y_value[i];
aom->y_points[i][1] = film_grain->point_y_scaling[i];
}
aom->num_uv_points[0] = film_grain->num_cb_points;
for (int i = 0; i < film_grain->num_cb_points; i++) {
aom->uv_points[0][i][0] = film_grain->point_cb_value[i];
aom->uv_points[0][i][1] = film_grain->point_cb_scaling[i];
}
aom->num_uv_points[1] = film_grain->num_cr_points;
for (int i = 0; i < film_grain->num_cr_points; i++) {
aom->uv_points[1][i][0] = film_grain->point_cr_value[i];
aom->uv_points[1][i][1] = film_grain->point_cr_scaling[i];
}
for (int i = 0; i < 24; i++) {
aom->ar_coeffs_y[i] = film_grain->ar_coeffs_y_plus_128[i] - 128;
}
for (int i = 0; i < 25; i++) {
aom->ar_coeffs_uv[0][i] = film_grain->ar_coeffs_cb_plus_128[i] - 128;
aom->ar_coeffs_uv[1][i] = film_grain->ar_coeffs_cr_plus_128[i] - 128;
}
aom->uv_mult[0] = film_grain->cb_mult;
aom->uv_mult[1] = film_grain->cr_mult;
aom->uv_mult_luma[0] = film_grain->cb_luma_mult;
aom->uv_mult_luma[1] = film_grain->cr_luma_mult;
aom->uv_offset[0] = film_grain->cb_offset;
aom->uv_offset[1] = film_grain->cr_offset;
return 0;
}
static int set_output_frame(AVCodecContext *avctx, AVFrame *frame,
const AVPacket *pkt, int *got_frame)
{
AV1DecContext *s = avctx->priv_data;
const AVFrame *srcframe = s->cur_frame.tf.f;
int ret;
ret = av_frame_ref(frame, srcframe);
if (ret < 0)
return ret;
if (avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) {
ret = export_film_grain(avctx, frame);
if (ret < 0) {
av_frame_unref(frame);
return ret;
}
}
frame->pts = pkt->pts;
frame->pkt_dts = pkt->dts;
frame->pkt_size = pkt->size;
*got_frame = 1;
return 0;
}
static int update_reference_list(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
const AV1RawFrameHeader *header = s->raw_frame_header;
int ret;
for (int i = 0; i < AV1_NUM_REF_FRAMES; i++) {
if (header->refresh_frame_flags & (1 << i)) {
if (s->ref[i].tf.f->buf[0])
av1_frame_unref(avctx, &s->ref[i]);
if ((ret = av1_frame_ref(avctx, &s->ref[i], &s->cur_frame)) < 0) {
av_log(avctx, AV_LOG_ERROR,
"Failed to update frame %d in reference list\n", i);
return ret;
}
}
}
return 0;
}
static int get_current_frame(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
int ret;
if (s->cur_frame.tf.f->buf[0])
av1_frame_unref(avctx, &s->cur_frame);
ret = av1_frame_alloc(avctx, &s->cur_frame);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR,
"Failed to allocate space for current frame.\n");
return ret;
}
ret = init_tile_data(s);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to init tile data.\n");
return ret;
}
global_motion_params(s);
skip_mode_params(s);
coded_lossless_param(s);
load_grain_params(s);
return ret;
}
static int av1_decode_frame(AVCodecContext *avctx, void *frame,
int *got_frame, AVPacket *pkt)
{
AV1DecContext *s = avctx->priv_data;
AV1RawTileGroup *raw_tile_group = NULL;
int ret;
ret = ff_cbs_read_packet(s->cbc, &s->current_obu, pkt);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to read packet.\n");
goto end;
}
av_log(avctx, AV_LOG_DEBUG, "Total obu for this frame:%d.\n",
s->current_obu.nb_units);
for (int i = 0; i < s->current_obu.nb_units; i++) {
CodedBitstreamUnit *unit = &s->current_obu.units[i];
AV1RawOBU *obu = unit->content;
const AV1RawOBUHeader *header;
if (!obu)
continue;
header = &obu->header;
av_log(avctx, AV_LOG_DEBUG, "Obu idx:%d, obu type:%d.\n", i, unit->type);
switch (unit->type) {
case AV1_OBU_SEQUENCE_HEADER:
av_buffer_unref(&s->seq_ref);
s->seq_ref = av_buffer_ref(unit->content_ref);
if (!s->seq_ref) {
ret = AVERROR(ENOMEM);
goto end;
}
s->raw_seq = &obu->obu.sequence_header;
ret = set_context_with_sequence(avctx, s->raw_seq);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to set context.\n");
s->raw_seq = NULL;
goto end;
}
if (s->pix_fmt == AV_PIX_FMT_NONE) {
ret = get_pixel_format(avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR,
"Failed to get pixel format.\n");
s->raw_seq = NULL;
goto end;
}
}
if (avctx->hwaccel && avctx->hwaccel->decode_params) {
ret = avctx->hwaccel->decode_params(avctx, unit->type, unit->data,
unit->data_size);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "HW accel decode params fail.\n");
s->raw_seq = NULL;
goto end;
}
}
break;
case AV1_OBU_REDUNDANT_FRAME_HEADER:
if (s->raw_frame_header)
break;
// fall-through
case AV1_OBU_FRAME:
case AV1_OBU_FRAME_HEADER:
if (!s->raw_seq) {
av_log(avctx, AV_LOG_ERROR, "Missing Sequence Header.\n");
ret = AVERROR_INVALIDDATA;
goto end;
}
av_buffer_unref(&s->header_ref);
s->header_ref = av_buffer_ref(unit->content_ref);
if (!s->header_ref) {
ret = AVERROR(ENOMEM);
goto end;
}
if (unit->type == AV1_OBU_FRAME)
s->raw_frame_header = &obu->obu.frame.header;
else
s->raw_frame_header = &obu->obu.frame_header;
if (s->raw_frame_header->show_existing_frame) {
if (s->cur_frame.tf.f->buf[0])
av1_frame_unref(avctx, &s->cur_frame);
ret = av1_frame_ref(avctx, &s->cur_frame,
&s->ref[s->raw_frame_header->frame_to_show_map_idx]);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to get reference frame.\n");
goto end;
}
ret = update_reference_list(avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to update reference list.\n");
goto end;
}
ret = set_output_frame(avctx, frame, pkt, got_frame);
if (ret < 0)
av_log(avctx, AV_LOG_ERROR, "Set output frame error.\n");
s->raw_frame_header = NULL;
goto end;
}
ret = get_current_frame(avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Get current frame error\n");
goto end;
}
s->cur_frame.spatial_id = header->spatial_id;
s->cur_frame.temporal_id = header->temporal_id;
if (avctx->hwaccel) {
ret = avctx->hwaccel->start_frame(avctx, unit->data,
unit->data_size);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "HW accel start frame fail.\n");
goto end;
}
}
if (unit->type != AV1_OBU_FRAME)
break;
// fall-through
case AV1_OBU_TILE_GROUP:
if (!s->raw_frame_header) {
av_log(avctx, AV_LOG_ERROR, "Missing Frame Header.\n");
ret = AVERROR_INVALIDDATA;
goto end;
}
if (unit->type == AV1_OBU_FRAME)
raw_tile_group = &obu->obu.frame.tile_group;
else
raw_tile_group = &obu->obu.tile_group;
ret = get_tiles_info(avctx, raw_tile_group);
if (ret < 0)
goto end;
if (avctx->hwaccel) {
ret = avctx->hwaccel->decode_slice(avctx,
raw_tile_group->tile_data.data,
raw_tile_group->tile_data.data_size);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR,
"HW accel decode slice fail.\n");
goto end;
}
}
break;
case AV1_OBU_TILE_LIST:
case AV1_OBU_TEMPORAL_DELIMITER:
case AV1_OBU_PADDING:
case AV1_OBU_METADATA:
break;
default:
av_log(avctx, AV_LOG_DEBUG,
"Unknown obu type: %d (%"SIZE_SPECIFIER" bits).\n",
unit->type, unit->data_size);
}
if (raw_tile_group && (s->tile_num == raw_tile_group->tg_end + 1)) {
if (avctx->hwaccel) {
ret = avctx->hwaccel->end_frame(avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "HW accel end frame fail.\n");
goto end;
}
}
ret = update_reference_list(avctx);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Failed to update reference list.\n");
goto end;
}
if (s->raw_frame_header->show_frame) {
ret = set_output_frame(avctx, frame, pkt, got_frame);
if (ret < 0) {
av_log(avctx, AV_LOG_ERROR, "Set output frame error\n");
goto end;
}
}
raw_tile_group = NULL;
s->raw_frame_header = NULL;
}
}
end:
ff_cbs_fragment_reset(&s->current_obu);
if (ret < 0)
s->raw_frame_header = NULL;
return ret;
}
static void av1_decode_flush(AVCodecContext *avctx)
{
AV1DecContext *s = avctx->priv_data;
for (int i = 0; i < FF_ARRAY_ELEMS(s->ref); i++)
av1_frame_unref(avctx, &s->ref[i]);
av1_frame_unref(avctx, &s->cur_frame);
s->raw_frame_header = NULL;
s->raw_seq = NULL;
ff_cbs_flush(s->cbc);
}
AVCodec ff_av1_decoder = {
.name = "av1",
.long_name = NULL_IF_CONFIG_SMALL("Alliance for Open Media AV1"),
.type = AVMEDIA_TYPE_VIDEO,
.id = AV_CODEC_ID_AV1,
.priv_data_size = sizeof(AV1DecContext),
.init = av1_decode_init,
.close = av1_decode_free,
.decode = av1_decode_frame,
.capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_AVOID_PROBING,
.caps_internal = FF_CODEC_CAP_INIT_THREADSAFE |
FF_CODEC_CAP_INIT_CLEANUP |
FF_CODEC_CAP_SETS_PKT_DTS,
.flush = av1_decode_flush,
.profiles = NULL_IF_CONFIG_SMALL(ff_av1_profiles),
.hw_configs = (const AVCodecHWConfigInternal * []) {
#if CONFIG_AV1_DXVA2_HWACCEL
HWACCEL_DXVA2(av1),
#endif
#if CONFIG_AV1_D3D11VA_HWACCEL
HWACCEL_D3D11VA(av1),
#endif
#if CONFIG_AV1_D3D11VA2_HWACCEL
HWACCEL_D3D11VA2(av1),
#endif
#if CONFIG_AV1_NVDEC_HWACCEL
HWACCEL_NVDEC(av1),
#endif
#if CONFIG_AV1_VAAPI_HWACCEL
HWACCEL_VAAPI(av1),
#endif
NULL
},
};