/* * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include #include "libavutil/avassert.h" #include "libavutil/common.h" #include "libavutil/mem.h" #include "libavutil/pixdesc.h" #include "libavutil/opt.h" #include "atsc_a53.h" #include "avcodec.h" #include "cbs.h" #include "cbs_h264.h" #include "codec_internal.h" #include "h264.h" #include "h264_levels.h" #include "h2645data.h" #include "vaapi_encode.h" #include "version.h" enum { SEI_TIMING = 0x01, SEI_IDENTIFIER = 0x02, SEI_RECOVERY_POINT = 0x04, SEI_A53_CC = 0x08, }; // Random (version 4) ISO 11578 UUID. static const uint8_t vaapi_encode_h264_sei_identifier_uuid[16] = { 0x59, 0x94, 0x8b, 0x28, 0x11, 0xec, 0x45, 0xaf, 0x96, 0x75, 0x19, 0xd4, 0x1f, 0xea, 0xa9, 0x4d, }; typedef struct VAAPIEncodeH264Picture { int frame_num; int pic_order_cnt; int64_t last_idr_frame; uint16_t idr_pic_id; int primary_pic_type; int slice_type; int cpb_delay; int dpb_delay; } VAAPIEncodeH264Picture; typedef struct VAAPIEncodeH264Context { VAAPIEncodeContext common; // User options. int qp; int quality; int coder; int aud; int sei; int profile; int level; // Derived settings. int mb_width; int mb_height; int fixed_qp_idr; int fixed_qp_p; int fixed_qp_b; int dpb_frames; // Writer structures. CodedBitstreamContext *cbc; CodedBitstreamFragment current_access_unit; H264RawAUD raw_aud; H264RawSPS raw_sps; H264RawPPS raw_pps; H264RawSlice raw_slice; H264RawSEIBufferingPeriod sei_buffering_period; H264RawSEIPicTiming sei_pic_timing; H264RawSEIRecoveryPoint sei_recovery_point; SEIRawUserDataUnregistered sei_identifier; char *sei_identifier_string; SEIRawUserDataRegistered sei_a53cc; void *sei_a53cc_data; int aud_needed; int sei_needed; int sei_cbr_workaround_needed; } VAAPIEncodeH264Context; static int vaapi_encode_h264_write_access_unit(AVCodecContext *avctx, char *data, size_t *data_len, CodedBitstreamFragment *au) { VAAPIEncodeH264Context *priv = avctx->priv_data; int err; err = ff_cbs_write_fragment_data(priv->cbc, au); if (err < 0) { av_log(avctx, AV_LOG_ERROR, "Failed to write packed header.\n"); return err; } if (*data_len < 8 * au->data_size - au->data_bit_padding) { av_log(avctx, AV_LOG_ERROR, "Access unit too large: " "%zu < %zu.\n", *data_len, 8 * au->data_size - au->data_bit_padding); return AVERROR(ENOSPC); } memcpy(data, au->data, au->data_size); *data_len = 8 * au->data_size - au->data_bit_padding; return 0; } static int vaapi_encode_h264_add_nal(AVCodecContext *avctx, CodedBitstreamFragment *au, void *nal_unit) { H264RawNALUnitHeader *header = nal_unit; int err; err = ff_cbs_insert_unit_content(au, -1, header->nal_unit_type, nal_unit, NULL); if (err < 0) { av_log(avctx, AV_LOG_ERROR, "Failed to add NAL unit: " "type = %d.\n", header->nal_unit_type); return err; } return 0; } static int vaapi_encode_h264_write_sequence_header(AVCodecContext *avctx, char *data, size_t *data_len) { VAAPIEncodeH264Context *priv = avctx->priv_data; CodedBitstreamFragment *au = &priv->current_access_unit; int err; if (priv->aud_needed) { err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_aud); if (err < 0) goto fail; priv->aud_needed = 0; } err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_sps); if (err < 0) goto fail; err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_pps); if (err < 0) goto fail; err = vaapi_encode_h264_write_access_unit(avctx, data, data_len, au); fail: ff_cbs_fragment_reset(au); return err; } static int vaapi_encode_h264_write_slice_header(AVCodecContext *avctx, VAAPIEncodePicture *pic, VAAPIEncodeSlice *slice, char *data, size_t *data_len) { VAAPIEncodeH264Context *priv = avctx->priv_data; CodedBitstreamFragment *au = &priv->current_access_unit; int err; if (priv->aud_needed) { err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_aud); if (err < 0) goto fail; priv->aud_needed = 0; } err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_slice); if (err < 0) goto fail; err = vaapi_encode_h264_write_access_unit(avctx, data, data_len, au); fail: ff_cbs_fragment_reset(au); return err; } static int vaapi_encode_h264_write_extra_header(AVCodecContext *avctx, VAAPIEncodePicture *pic, int index, int *type, char *data, size_t *data_len) { VAAPIEncodeH264Context *priv = avctx->priv_data; CodedBitstreamFragment *au = &priv->current_access_unit; int err; if (priv->sei_needed) { if (priv->aud_needed) { err = vaapi_encode_h264_add_nal(avctx, au, &priv->raw_aud); if (err < 0) goto fail; priv->aud_needed = 0; } if (priv->sei_needed & SEI_IDENTIFIER) { err = ff_cbs_sei_add_message(priv->cbc, au, 1, SEI_TYPE_USER_DATA_UNREGISTERED, &priv->sei_identifier, NULL); if (err < 0) goto fail; } if (priv->sei_needed & SEI_TIMING) { if (pic->type == PICTURE_TYPE_IDR) { err = ff_cbs_sei_add_message(priv->cbc, au, 1, SEI_TYPE_BUFFERING_PERIOD, &priv->sei_buffering_period, NULL); if (err < 0) goto fail; } err = ff_cbs_sei_add_message(priv->cbc, au, 1, SEI_TYPE_PIC_TIMING, &priv->sei_pic_timing, NULL); if (err < 0) goto fail; } if (priv->sei_needed & SEI_RECOVERY_POINT) { err = ff_cbs_sei_add_message(priv->cbc, au, 1, SEI_TYPE_RECOVERY_POINT, &priv->sei_recovery_point, NULL); if (err < 0) goto fail; } if (priv->sei_needed & SEI_A53_CC) { err = ff_cbs_sei_add_message(priv->cbc, au, 1, SEI_TYPE_USER_DATA_REGISTERED_ITU_T_T35, &priv->sei_a53cc, NULL); if (err < 0) goto fail; } priv->sei_needed = 0; err = vaapi_encode_h264_write_access_unit(avctx, data, data_len, au); if (err < 0) goto fail; ff_cbs_fragment_reset(au); *type = VAEncPackedHeaderRawData; return 0; #if !CONFIG_VAAPI_1 } else if (priv->sei_cbr_workaround_needed) { // Insert a zero-length header using the old SEI type. This is // required to avoid triggering broken behaviour on Intel platforms // in CBR mode where an invalid SEI message is generated by the // driver and inserted into the stream. *data_len = 0; *type = VAEncPackedHeaderH264_SEI; priv->sei_cbr_workaround_needed = 0; return 0; #endif } else { return AVERROR_EOF; } fail: ff_cbs_fragment_reset(au); return err; } static int vaapi_encode_h264_init_sequence_params(AVCodecContext *avctx) { VAAPIEncodeContext *ctx = avctx->priv_data; VAAPIEncodeH264Context *priv = avctx->priv_data; H264RawSPS *sps = &priv->raw_sps; H264RawPPS *pps = &priv->raw_pps; VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params; VAEncPictureParameterBufferH264 *vpic = ctx->codec_picture_params; const AVPixFmtDescriptor *desc; int bit_depth; memset(sps, 0, sizeof(*sps)); memset(pps, 0, sizeof(*pps)); desc = av_pix_fmt_desc_get(priv->common.input_frames->sw_format); av_assert0(desc); if (desc->nb_components == 1 || desc->log2_chroma_w != 1 || desc->log2_chroma_h != 1) { av_log(avctx, AV_LOG_ERROR, "Chroma format of input pixel format " "%s is not supported.\n", desc->name); return AVERROR(EINVAL); } bit_depth = desc->comp[0].depth; sps->nal_unit_header.nal_ref_idc = 3; sps->nal_unit_header.nal_unit_type = H264_NAL_SPS; sps->profile_idc = avctx->profile & 0xff; if (avctx->profile == AV_PROFILE_H264_CONSTRAINED_BASELINE || avctx->profile == AV_PROFILE_H264_MAIN) sps->constraint_set1_flag = 1; if (avctx->profile == AV_PROFILE_H264_HIGH || avctx->profile == AV_PROFILE_H264_HIGH_10) sps->constraint_set3_flag = ctx->gop_size == 1; if (avctx->profile == AV_PROFILE_H264_MAIN || avctx->profile == AV_PROFILE_H264_HIGH || avctx->profile == AV_PROFILE_H264_HIGH_10) { sps->constraint_set4_flag = 1; sps->constraint_set5_flag = ctx->b_per_p == 0; } if (ctx->gop_size == 1) priv->dpb_frames = 0; else priv->dpb_frames = 1 + ctx->max_b_depth; if (avctx->level != AV_LEVEL_UNKNOWN) { sps->level_idc = avctx->level; } else { const H264LevelDescriptor *level; int framerate; if (avctx->framerate.num > 0 && avctx->framerate.den > 0) framerate = avctx->framerate.num / avctx->framerate.den; else framerate = 0; level = ff_h264_guess_level(sps->profile_idc, avctx->bit_rate, framerate, priv->mb_width * 16, priv->mb_height * 16, priv->dpb_frames); if (level) { av_log(avctx, AV_LOG_VERBOSE, "Using level %s.\n", level->name); if (level->constraint_set3_flag) sps->constraint_set3_flag = 1; sps->level_idc = level->level_idc; } else { av_log(avctx, AV_LOG_WARNING, "Stream will not conform " "to any level: using level 6.2.\n"); sps->level_idc = 62; } } sps->seq_parameter_set_id = 0; sps->chroma_format_idc = 1; sps->bit_depth_luma_minus8 = bit_depth - 8; sps->bit_depth_chroma_minus8 = bit_depth - 8; sps->log2_max_frame_num_minus4 = 4; sps->pic_order_cnt_type = ctx->max_b_depth ? 0 : 2; if (sps->pic_order_cnt_type == 0) { sps->log2_max_pic_order_cnt_lsb_minus4 = 4; } sps->max_num_ref_frames = priv->dpb_frames; sps->pic_width_in_mbs_minus1 = priv->mb_width - 1; sps->pic_height_in_map_units_minus1 = priv->mb_height - 1; sps->frame_mbs_only_flag = 1; sps->direct_8x8_inference_flag = 1; if (avctx->width != 16 * priv->mb_width || avctx->height != 16 * priv->mb_height) { sps->frame_cropping_flag = 1; sps->frame_crop_left_offset = 0; sps->frame_crop_right_offset = (16 * priv->mb_width - avctx->width) / 2; sps->frame_crop_top_offset = 0; sps->frame_crop_bottom_offset = (16 * priv->mb_height - avctx->height) / 2; } else { sps->frame_cropping_flag = 0; } sps->vui_parameters_present_flag = 1; if (avctx->sample_aspect_ratio.num != 0 && avctx->sample_aspect_ratio.den != 0) { int num, den, i; av_reduce(&num, &den, avctx->sample_aspect_ratio.num, avctx->sample_aspect_ratio.den, 65535); for (i = 0; i < FF_ARRAY_ELEMS(ff_h2645_pixel_aspect); i++) { if (num == ff_h2645_pixel_aspect[i].num && den == ff_h2645_pixel_aspect[i].den) { sps->vui.aspect_ratio_idc = i; break; } } if (i >= FF_ARRAY_ELEMS(ff_h2645_pixel_aspect)) { sps->vui.aspect_ratio_idc = 255; sps->vui.sar_width = num; sps->vui.sar_height = den; } sps->vui.aspect_ratio_info_present_flag = 1; } // Unspecified video format, from table E-2. sps->vui.video_format = 5; sps->vui.video_full_range_flag = avctx->color_range == AVCOL_RANGE_JPEG; sps->vui.colour_primaries = avctx->color_primaries; sps->vui.transfer_characteristics = avctx->color_trc; sps->vui.matrix_coefficients = avctx->colorspace; if (avctx->color_primaries != AVCOL_PRI_UNSPECIFIED || avctx->color_trc != AVCOL_TRC_UNSPECIFIED || avctx->colorspace != AVCOL_SPC_UNSPECIFIED) sps->vui.colour_description_present_flag = 1; if (avctx->color_range != AVCOL_RANGE_UNSPECIFIED || sps->vui.colour_description_present_flag) sps->vui.video_signal_type_present_flag = 1; if (avctx->chroma_sample_location != AVCHROMA_LOC_UNSPECIFIED) { sps->vui.chroma_loc_info_present_flag = 1; sps->vui.chroma_sample_loc_type_top_field = sps->vui.chroma_sample_loc_type_bottom_field = avctx->chroma_sample_location - 1; } sps->vui.timing_info_present_flag = 1; if (avctx->framerate.num > 0 && avctx->framerate.den > 0) { sps->vui.num_units_in_tick = avctx->framerate.den; sps->vui.time_scale = 2 * avctx->framerate.num; sps->vui.fixed_frame_rate_flag = 1; } else { sps->vui.num_units_in_tick = avctx->time_base.num; sps->vui.time_scale = 2 * avctx->time_base.den; sps->vui.fixed_frame_rate_flag = 0; } if (priv->sei & SEI_TIMING) { H264RawHRD *hrd = &sps->vui.nal_hrd_parameters; H264RawSEIBufferingPeriod *bp = &priv->sei_buffering_period; sps->vui.nal_hrd_parameters_present_flag = 1; hrd->cpb_cnt_minus1 = 0; // Try to scale these to a sensible range so that the // golomb encode of the value is not overlong. hrd->bit_rate_scale = av_clip_uintp2(av_log2(ctx->va_bit_rate) - 15 - 6, 4); hrd->bit_rate_value_minus1[0] = (ctx->va_bit_rate >> hrd->bit_rate_scale + 6) - 1; hrd->cpb_size_scale = av_clip_uintp2(av_log2(ctx->hrd_params.buffer_size) - 15 - 4, 4); hrd->cpb_size_value_minus1[0] = (ctx->hrd_params.buffer_size >> hrd->cpb_size_scale + 4) - 1; // CBR mode as defined for the HRD cannot be achieved without filler // data, so this flag cannot be set even with VAAPI CBR modes. hrd->cbr_flag[0] = 0; hrd->initial_cpb_removal_delay_length_minus1 = 23; hrd->cpb_removal_delay_length_minus1 = 23; hrd->dpb_output_delay_length_minus1 = 7; hrd->time_offset_length = 0; bp->seq_parameter_set_id = sps->seq_parameter_set_id; // This calculation can easily overflow 32 bits. bp->nal.initial_cpb_removal_delay[0] = 90000 * (uint64_t)ctx->hrd_params.initial_buffer_fullness / ctx->hrd_params.buffer_size; bp->nal.initial_cpb_removal_delay_offset[0] = 0; } else { sps->vui.nal_hrd_parameters_present_flag = 0; sps->vui.low_delay_hrd_flag = 1 - sps->vui.fixed_frame_rate_flag; } sps->vui.bitstream_restriction_flag = 1; sps->vui.motion_vectors_over_pic_boundaries_flag = 1; sps->vui.log2_max_mv_length_horizontal = 15; sps->vui.log2_max_mv_length_vertical = 15; sps->vui.max_num_reorder_frames = ctx->max_b_depth; sps->vui.max_dec_frame_buffering = ctx->max_b_depth + 1; pps->nal_unit_header.nal_ref_idc = 3; pps->nal_unit_header.nal_unit_type = H264_NAL_PPS; pps->pic_parameter_set_id = 0; pps->seq_parameter_set_id = 0; pps->entropy_coding_mode_flag = !(sps->profile_idc == AV_PROFILE_H264_BASELINE || sps->profile_idc == AV_PROFILE_H264_EXTENDED || sps->profile_idc == AV_PROFILE_H264_CAVLC_444); if (!priv->coder && pps->entropy_coding_mode_flag) pps->entropy_coding_mode_flag = 0; pps->num_ref_idx_l0_default_active_minus1 = 0; pps->num_ref_idx_l1_default_active_minus1 = 0; pps->pic_init_qp_minus26 = priv->fixed_qp_idr - 26; if (sps->profile_idc == AV_PROFILE_H264_BASELINE || sps->profile_idc == AV_PROFILE_H264_EXTENDED || sps->profile_idc == AV_PROFILE_H264_MAIN) { pps->more_rbsp_data = 0; } else { pps->more_rbsp_data = 1; pps->transform_8x8_mode_flag = 1; } *vseq = (VAEncSequenceParameterBufferH264) { .seq_parameter_set_id = sps->seq_parameter_set_id, .level_idc = sps->level_idc, .intra_period = ctx->gop_size, .intra_idr_period = ctx->gop_size, .ip_period = ctx->b_per_p + 1, .bits_per_second = ctx->va_bit_rate, .max_num_ref_frames = sps->max_num_ref_frames, .picture_width_in_mbs = sps->pic_width_in_mbs_minus1 + 1, .picture_height_in_mbs = sps->pic_height_in_map_units_minus1 + 1, .seq_fields.bits = { .chroma_format_idc = sps->chroma_format_idc, .frame_mbs_only_flag = sps->frame_mbs_only_flag, .mb_adaptive_frame_field_flag = sps->mb_adaptive_frame_field_flag, .seq_scaling_matrix_present_flag = sps->seq_scaling_matrix_present_flag, .direct_8x8_inference_flag = sps->direct_8x8_inference_flag, .log2_max_frame_num_minus4 = sps->log2_max_frame_num_minus4, .pic_order_cnt_type = sps->pic_order_cnt_type, .log2_max_pic_order_cnt_lsb_minus4 = sps->log2_max_pic_order_cnt_lsb_minus4, .delta_pic_order_always_zero_flag = sps->delta_pic_order_always_zero_flag, }, .bit_depth_luma_minus8 = sps->bit_depth_luma_minus8, .bit_depth_chroma_minus8 = sps->bit_depth_chroma_minus8, .frame_cropping_flag = sps->frame_cropping_flag, .frame_crop_left_offset = sps->frame_crop_left_offset, .frame_crop_right_offset = sps->frame_crop_right_offset, .frame_crop_top_offset = sps->frame_crop_top_offset, .frame_crop_bottom_offset = sps->frame_crop_bottom_offset, .vui_parameters_present_flag = sps->vui_parameters_present_flag, .vui_fields.bits = { .aspect_ratio_info_present_flag = sps->vui.aspect_ratio_info_present_flag, .timing_info_present_flag = sps->vui.timing_info_present_flag, .bitstream_restriction_flag = sps->vui.bitstream_restriction_flag, .log2_max_mv_length_horizontal = sps->vui.log2_max_mv_length_horizontal, .log2_max_mv_length_vertical = sps->vui.log2_max_mv_length_vertical, }, .aspect_ratio_idc = sps->vui.aspect_ratio_idc, .sar_width = sps->vui.sar_width, .sar_height = sps->vui.sar_height, .num_units_in_tick = sps->vui.num_units_in_tick, .time_scale = sps->vui.time_scale, }; *vpic = (VAEncPictureParameterBufferH264) { .CurrPic = { .picture_id = VA_INVALID_ID, .flags = VA_PICTURE_H264_INVALID, }, .coded_buf = VA_INVALID_ID, .pic_parameter_set_id = pps->pic_parameter_set_id, .seq_parameter_set_id = pps->seq_parameter_set_id, .pic_init_qp = pps->pic_init_qp_minus26 + 26, .num_ref_idx_l0_active_minus1 = pps->num_ref_idx_l0_default_active_minus1, .num_ref_idx_l1_active_minus1 = pps->num_ref_idx_l1_default_active_minus1, .chroma_qp_index_offset = pps->chroma_qp_index_offset, .second_chroma_qp_index_offset = pps->second_chroma_qp_index_offset, .pic_fields.bits = { .entropy_coding_mode_flag = pps->entropy_coding_mode_flag, .weighted_pred_flag = pps->weighted_pred_flag, .weighted_bipred_idc = pps->weighted_bipred_idc, .constrained_intra_pred_flag = pps->constrained_intra_pred_flag, .transform_8x8_mode_flag = pps->transform_8x8_mode_flag, .deblocking_filter_control_present_flag = pps->deblocking_filter_control_present_flag, .redundant_pic_cnt_present_flag = pps->redundant_pic_cnt_present_flag, .pic_order_present_flag = pps->bottom_field_pic_order_in_frame_present_flag, .pic_scaling_matrix_present_flag = pps->pic_scaling_matrix_present_flag, }, }; return 0; } static int vaapi_encode_h264_init_picture_params(AVCodecContext *avctx, VAAPIEncodePicture *pic) { VAAPIEncodeContext *ctx = avctx->priv_data; VAAPIEncodeH264Context *priv = avctx->priv_data; VAAPIEncodeH264Picture *hpic = pic->priv_data; VAAPIEncodePicture *prev = pic->prev; VAAPIEncodeH264Picture *hprev = prev ? prev->priv_data : NULL; VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params; int i, j = 0; if (pic->type == PICTURE_TYPE_IDR) { av_assert0(pic->display_order == pic->encode_order); hpic->frame_num = 0; hpic->last_idr_frame = pic->display_order; hpic->idr_pic_id = hprev ? hprev->idr_pic_id + 1 : 0; hpic->primary_pic_type = 0; hpic->slice_type = 7; } else { av_assert0(prev); hpic->frame_num = hprev->frame_num + prev->is_reference; hpic->last_idr_frame = hprev->last_idr_frame; hpic->idr_pic_id = hprev->idr_pic_id; if (pic->type == PICTURE_TYPE_I) { hpic->slice_type = 7; hpic->primary_pic_type = 0; } else if (pic->type == PICTURE_TYPE_P) { hpic->slice_type = 5; hpic->primary_pic_type = 1; } else { hpic->slice_type = 6; hpic->primary_pic_type = 2; } } hpic->pic_order_cnt = pic->display_order - hpic->last_idr_frame; if (priv->raw_sps.pic_order_cnt_type == 2) { hpic->pic_order_cnt *= 2; } hpic->dpb_delay = pic->display_order - pic->encode_order + ctx->max_b_depth; hpic->cpb_delay = pic->encode_order - hpic->last_idr_frame; if (priv->aud) { priv->aud_needed = 1; priv->raw_aud = (H264RawAUD) { .nal_unit_header = { .nal_unit_type = H264_NAL_AUD, }, .primary_pic_type = hpic->primary_pic_type, }; } else { priv->aud_needed = 0; } priv->sei_needed = 0; if (priv->sei & SEI_IDENTIFIER && pic->encode_order == 0) priv->sei_needed |= SEI_IDENTIFIER; #if !CONFIG_VAAPI_1 if (ctx->va_rc_mode == VA_RC_CBR) priv->sei_cbr_workaround_needed = 1; #endif if (priv->sei & SEI_TIMING) { priv->sei_pic_timing = (H264RawSEIPicTiming) { .cpb_removal_delay = 2 * hpic->cpb_delay, .dpb_output_delay = 2 * hpic->dpb_delay, }; priv->sei_needed |= SEI_TIMING; } if (priv->sei & SEI_RECOVERY_POINT && pic->type == PICTURE_TYPE_I) { priv->sei_recovery_point = (H264RawSEIRecoveryPoint) { .recovery_frame_cnt = 0, .exact_match_flag = 1, .broken_link_flag = ctx->b_per_p > 0, }; priv->sei_needed |= SEI_RECOVERY_POINT; } if (priv->sei & SEI_A53_CC) { int err; size_t sei_a53cc_len; av_freep(&priv->sei_a53cc_data); err = ff_alloc_a53_sei(pic->input_image, 0, &priv->sei_a53cc_data, &sei_a53cc_len); if (err < 0) return err; if (priv->sei_a53cc_data != NULL) { priv->sei_a53cc.itu_t_t35_country_code = 181; priv->sei_a53cc.data = (uint8_t *)priv->sei_a53cc_data + 1; priv->sei_a53cc.data_length = sei_a53cc_len - 1; priv->sei_needed |= SEI_A53_CC; } } vpic->CurrPic = (VAPictureH264) { .picture_id = pic->recon_surface, .frame_idx = hpic->frame_num, .flags = 0, .TopFieldOrderCnt = hpic->pic_order_cnt, .BottomFieldOrderCnt = hpic->pic_order_cnt, }; for (int k = 0; k < MAX_REFERENCE_LIST_NUM; k++) { for (i = 0; i < pic->nb_refs[k]; i++) { VAAPIEncodePicture *ref = pic->refs[k][i]; VAAPIEncodeH264Picture *href; av_assert0(ref && ref->encode_order < pic->encode_order); href = ref->priv_data; vpic->ReferenceFrames[j++] = (VAPictureH264) { .picture_id = ref->recon_surface, .frame_idx = href->frame_num, .flags = VA_PICTURE_H264_SHORT_TERM_REFERENCE, .TopFieldOrderCnt = href->pic_order_cnt, .BottomFieldOrderCnt = href->pic_order_cnt, }; } } for (; j < FF_ARRAY_ELEMS(vpic->ReferenceFrames); j++) { vpic->ReferenceFrames[j] = (VAPictureH264) { .picture_id = VA_INVALID_ID, .flags = VA_PICTURE_H264_INVALID, }; } vpic->coded_buf = pic->output_buffer; vpic->frame_num = hpic->frame_num; vpic->pic_fields.bits.idr_pic_flag = (pic->type == PICTURE_TYPE_IDR); vpic->pic_fields.bits.reference_pic_flag = (pic->type != PICTURE_TYPE_B); return 0; } static void vaapi_encode_h264_default_ref_pic_list(AVCodecContext *avctx, VAAPIEncodePicture *pic, VAAPIEncodePicture **rpl0, VAAPIEncodePicture **rpl1, int *rpl_size) { VAAPIEncodePicture *prev; VAAPIEncodeH264Picture *hp, *hn, *hc; int i, j, n = 0; prev = pic->prev; av_assert0(prev); hp = pic->priv_data; for (i = 0; i < pic->prev->nb_dpb_pics; i++) { hn = prev->dpb[i]->priv_data; av_assert0(hn->frame_num < hp->frame_num); if (pic->type == PICTURE_TYPE_P) { for (j = n; j > 0; j--) { hc = rpl0[j - 1]->priv_data; av_assert0(hc->frame_num != hn->frame_num); if (hc->frame_num > hn->frame_num) break; rpl0[j] = rpl0[j - 1]; } rpl0[j] = prev->dpb[i]; } else if (pic->type == PICTURE_TYPE_B) { for (j = n; j > 0; j--) { hc = rpl0[j - 1]->priv_data; av_assert0(hc->pic_order_cnt != hp->pic_order_cnt); if (hc->pic_order_cnt < hp->pic_order_cnt) { if (hn->pic_order_cnt > hp->pic_order_cnt || hn->pic_order_cnt < hc->pic_order_cnt) break; } else { if (hn->pic_order_cnt > hc->pic_order_cnt) break; } rpl0[j] = rpl0[j - 1]; } rpl0[j] = prev->dpb[i]; for (j = n; j > 0; j--) { hc = rpl1[j - 1]->priv_data; av_assert0(hc->pic_order_cnt != hp->pic_order_cnt); if (hc->pic_order_cnt > hp->pic_order_cnt) { if (hn->pic_order_cnt < hp->pic_order_cnt || hn->pic_order_cnt > hc->pic_order_cnt) break; } else { if (hn->pic_order_cnt < hc->pic_order_cnt) break; } rpl1[j] = rpl1[j - 1]; } rpl1[j] = prev->dpb[i]; } ++n; } if (pic->type == PICTURE_TYPE_B) { for (i = 0; i < n; i++) { if (rpl0[i] != rpl1[i]) break; } if (i == n) FFSWAP(VAAPIEncodePicture*, rpl1[0], rpl1[1]); } if (pic->type == PICTURE_TYPE_P || pic->type == PICTURE_TYPE_B) { av_log(avctx, AV_LOG_DEBUG, "Default RefPicList0 for fn=%d/poc=%d:", hp->frame_num, hp->pic_order_cnt); for (i = 0; i < n; i++) { hn = rpl0[i]->priv_data; av_log(avctx, AV_LOG_DEBUG, " fn=%d/poc=%d", hn->frame_num, hn->pic_order_cnt); } av_log(avctx, AV_LOG_DEBUG, "\n"); } if (pic->type == PICTURE_TYPE_B) { av_log(avctx, AV_LOG_DEBUG, "Default RefPicList1 for fn=%d/poc=%d:", hp->frame_num, hp->pic_order_cnt); for (i = 0; i < n; i++) { hn = rpl1[i]->priv_data; av_log(avctx, AV_LOG_DEBUG, " fn=%d/poc=%d", hn->frame_num, hn->pic_order_cnt); } av_log(avctx, AV_LOG_DEBUG, "\n"); } *rpl_size = n; } static int vaapi_encode_h264_init_slice_params(AVCodecContext *avctx, VAAPIEncodePicture *pic, VAAPIEncodeSlice *slice) { VAAPIEncodeH264Context *priv = avctx->priv_data; VAAPIEncodeH264Picture *hpic = pic->priv_data; VAAPIEncodePicture *prev = pic->prev; H264RawSPS *sps = &priv->raw_sps; H264RawPPS *pps = &priv->raw_pps; H264RawSliceHeader *sh = &priv->raw_slice.header; VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params; VAEncSliceParameterBufferH264 *vslice = slice->codec_slice_params; int i, j; if (pic->type == PICTURE_TYPE_IDR) { sh->nal_unit_header.nal_unit_type = H264_NAL_IDR_SLICE; sh->nal_unit_header.nal_ref_idc = 3; } else { sh->nal_unit_header.nal_unit_type = H264_NAL_SLICE; sh->nal_unit_header.nal_ref_idc = pic->is_reference; } sh->first_mb_in_slice = slice->block_start; sh->slice_type = hpic->slice_type; sh->pic_parameter_set_id = pps->pic_parameter_set_id; sh->frame_num = hpic->frame_num & ((1 << (4 + sps->log2_max_frame_num_minus4)) - 1); sh->idr_pic_id = hpic->idr_pic_id; sh->pic_order_cnt_lsb = hpic->pic_order_cnt & ((1 << (4 + sps->log2_max_pic_order_cnt_lsb_minus4)) - 1); sh->direct_spatial_mv_pred_flag = 1; if (pic->type == PICTURE_TYPE_B) sh->slice_qp_delta = priv->fixed_qp_b - (pps->pic_init_qp_minus26 + 26); else if (pic->type == PICTURE_TYPE_P) sh->slice_qp_delta = priv->fixed_qp_p - (pps->pic_init_qp_minus26 + 26); else sh->slice_qp_delta = priv->fixed_qp_idr - (pps->pic_init_qp_minus26 + 26); if (pic->is_reference && pic->type != PICTURE_TYPE_IDR) { VAAPIEncodePicture *discard_list[MAX_DPB_SIZE]; int discard = 0, keep = 0; // Discard everything which is in the DPB of the previous frame but // not in the DPB of this one. for (i = 0; i < prev->nb_dpb_pics; i++) { for (j = 0; j < pic->nb_dpb_pics; j++) { if (prev->dpb[i] == pic->dpb[j]) break; } if (j == pic->nb_dpb_pics) { discard_list[discard] = prev->dpb[i]; ++discard; } else { ++keep; } } av_assert0(keep <= priv->dpb_frames); if (discard == 0) { sh->adaptive_ref_pic_marking_mode_flag = 0; } else { sh->adaptive_ref_pic_marking_mode_flag = 1; for (i = 0; i < discard; i++) { VAAPIEncodeH264Picture *old = discard_list[i]->priv_data; av_assert0(old->frame_num < hpic->frame_num); sh->mmco[i].memory_management_control_operation = 1; sh->mmco[i].difference_of_pic_nums_minus1 = hpic->frame_num - old->frame_num - 1; } sh->mmco[i].memory_management_control_operation = 0; } } // If the intended references are not the first entries of RefPicListN // by default, use ref-pic-list-modification to move them there. if (pic->type == PICTURE_TYPE_P || pic->type == PICTURE_TYPE_B) { VAAPIEncodePicture *def_l0[MAX_DPB_SIZE], *def_l1[MAX_DPB_SIZE]; VAAPIEncodeH264Picture *href; int n; vaapi_encode_h264_default_ref_pic_list(avctx, pic, def_l0, def_l1, &n); if (pic->type == PICTURE_TYPE_P) { int need_rplm = 0; for (i = 0; i < pic->nb_refs[0]; i++) { av_assert0(pic->refs[0][i]); if (pic->refs[0][i] != def_l0[i]) need_rplm = 1; } sh->ref_pic_list_modification_flag_l0 = need_rplm; if (need_rplm) { int pic_num = hpic->frame_num; for (i = 0; i < pic->nb_refs[0]; i++) { href = pic->refs[0][i]->priv_data; av_assert0(href->frame_num != pic_num); if (href->frame_num < pic_num) { sh->rplm_l0[i].modification_of_pic_nums_idc = 0; sh->rplm_l0[i].abs_diff_pic_num_minus1 = pic_num - href->frame_num - 1; } else { sh->rplm_l0[i].modification_of_pic_nums_idc = 1; sh->rplm_l0[i].abs_diff_pic_num_minus1 = href->frame_num - pic_num - 1; } pic_num = href->frame_num; } sh->rplm_l0[i].modification_of_pic_nums_idc = 3; } } else { int need_rplm_l0 = 0, need_rplm_l1 = 0; int n0 = 0, n1 = 0; for (i = 0; i < pic->nb_refs[0]; i++) { av_assert0(pic->refs[0][i]); href = pic->refs[0][i]->priv_data; av_assert0(href->pic_order_cnt < hpic->pic_order_cnt); if (pic->refs[0][i] != def_l0[n0]) need_rplm_l0 = 1; ++n0; } for (i = 0; i < pic->nb_refs[1]; i++) { av_assert0(pic->refs[1][i]); href = pic->refs[1][i]->priv_data; av_assert0(href->pic_order_cnt > hpic->pic_order_cnt); if (pic->refs[1][i] != def_l1[n1]) need_rplm_l1 = 1; ++n1; } sh->ref_pic_list_modification_flag_l0 = need_rplm_l0; if (need_rplm_l0) { int pic_num = hpic->frame_num; for (i = j = 0; i < pic->nb_refs[0]; i++) { href = pic->refs[0][i]->priv_data; av_assert0(href->frame_num != pic_num); if (href->frame_num < pic_num) { sh->rplm_l0[j].modification_of_pic_nums_idc = 0; sh->rplm_l0[j].abs_diff_pic_num_minus1 = pic_num - href->frame_num - 1; } else { sh->rplm_l0[j].modification_of_pic_nums_idc = 1; sh->rplm_l0[j].abs_diff_pic_num_minus1 = href->frame_num - pic_num - 1; } pic_num = href->frame_num; ++j; } av_assert0(j == n0); sh->rplm_l0[j].modification_of_pic_nums_idc = 3; } sh->ref_pic_list_modification_flag_l1 = need_rplm_l1; if (need_rplm_l1) { int pic_num = hpic->frame_num; for (i = j = 0; i < pic->nb_refs[1]; i++) { href = pic->refs[1][i]->priv_data; av_assert0(href->frame_num != pic_num); if (href->frame_num < pic_num) { sh->rplm_l1[j].modification_of_pic_nums_idc = 0; sh->rplm_l1[j].abs_diff_pic_num_minus1 = pic_num - href->frame_num - 1; } else { sh->rplm_l1[j].modification_of_pic_nums_idc = 1; sh->rplm_l1[j].abs_diff_pic_num_minus1 = href->frame_num - pic_num - 1; } pic_num = href->frame_num; ++j; } av_assert0(j == n1); sh->rplm_l1[j].modification_of_pic_nums_idc = 3; } } } vslice->macroblock_address = slice->block_start; vslice->num_macroblocks = slice->block_size; vslice->macroblock_info = VA_INVALID_ID; vslice->slice_type = sh->slice_type % 5; vslice->pic_parameter_set_id = sh->pic_parameter_set_id; vslice->idr_pic_id = sh->idr_pic_id; vslice->pic_order_cnt_lsb = sh->pic_order_cnt_lsb; vslice->direct_spatial_mv_pred_flag = sh->direct_spatial_mv_pred_flag; for (i = 0; i < FF_ARRAY_ELEMS(vslice->RefPicList0); i++) { vslice->RefPicList0[i].picture_id = VA_INVALID_ID; vslice->RefPicList0[i].flags = VA_PICTURE_H264_INVALID; vslice->RefPicList1[i].picture_id = VA_INVALID_ID; vslice->RefPicList1[i].flags = VA_PICTURE_H264_INVALID; } if (pic->nb_refs[0]) { // Backward reference for P- or B-frame. av_assert0(pic->type == PICTURE_TYPE_P || pic->type == PICTURE_TYPE_B); vslice->RefPicList0[0] = vpic->ReferenceFrames[0]; } if (pic->nb_refs[1]) { // Forward reference for B-frame. av_assert0(pic->type == PICTURE_TYPE_B); vslice->RefPicList1[0] = vpic->ReferenceFrames[1]; } vslice->slice_qp_delta = sh->slice_qp_delta; return 0; } static av_cold int vaapi_encode_h264_configure(AVCodecContext *avctx) { VAAPIEncodeContext *ctx = avctx->priv_data; VAAPIEncodeH264Context *priv = avctx->priv_data; int err; err = ff_cbs_init(&priv->cbc, AV_CODEC_ID_H264, avctx); if (err < 0) return err; priv->mb_width = FFALIGN(avctx->width, 16) / 16; priv->mb_height = FFALIGN(avctx->height, 16) / 16; if (ctx->va_rc_mode == VA_RC_CQP) { priv->fixed_qp_p = av_clip(ctx->rc_quality, 1, 51); if (avctx->i_quant_factor > 0.0) priv->fixed_qp_idr = av_clip((avctx->i_quant_factor * priv->fixed_qp_p + avctx->i_quant_offset) + 0.5, 1, 51); else priv->fixed_qp_idr = priv->fixed_qp_p; if (avctx->b_quant_factor > 0.0) priv->fixed_qp_b = av_clip((avctx->b_quant_factor * priv->fixed_qp_p + avctx->b_quant_offset) + 0.5, 1, 51); else priv->fixed_qp_b = priv->fixed_qp_p; av_log(avctx, AV_LOG_DEBUG, "Using fixed QP = " "%d / %d / %d for IDR- / P- / B-frames.\n", priv->fixed_qp_idr, priv->fixed_qp_p, priv->fixed_qp_b); } else { // These still need to be set for pic_init_qp/slice_qp_delta. priv->fixed_qp_idr = 26; priv->fixed_qp_p = 26; priv->fixed_qp_b = 26; } if (!ctx->rc_mode->hrd) { // Timing SEI requires a mode respecting HRD parameters. priv->sei &= ~SEI_TIMING; } if (priv->sei & SEI_IDENTIFIER) { const char *lavc = LIBAVCODEC_IDENT; const char *vaapi = VA_VERSION_S; const char *driver; int len; memcpy(priv->sei_identifier.uuid_iso_iec_11578, vaapi_encode_h264_sei_identifier_uuid, sizeof(priv->sei_identifier.uuid_iso_iec_11578)); driver = vaQueryVendorString(ctx->hwctx->display); if (!driver) driver = "unknown driver"; len = snprintf(NULL, 0, "%s / VAAPI %s / %s", lavc, vaapi, driver); if (len >= 0) { priv->sei_identifier_string = av_malloc(len + 1); if (!priv->sei_identifier_string) return AVERROR(ENOMEM); snprintf(priv->sei_identifier_string, len + 1, "%s / VAAPI %s / %s", lavc, vaapi, driver); priv->sei_identifier.data = priv->sei_identifier_string; priv->sei_identifier.data_length = len + 1; } } ctx->roi_quant_range = 51 + 6 * (ctx->profile->depth - 8); return 0; } static const VAAPIEncodeProfile vaapi_encode_h264_profiles[] = { #if VA_CHECK_VERSION(1, 18, 0) { AV_PROFILE_H264_HIGH_10, 10, 3, 1, 1, VAProfileH264High10 }, #endif { AV_PROFILE_H264_HIGH, 8, 3, 1, 1, VAProfileH264High }, { AV_PROFILE_H264_MAIN, 8, 3, 1, 1, VAProfileH264Main }, { AV_PROFILE_H264_CONSTRAINED_BASELINE, 8, 3, 1, 1, VAProfileH264ConstrainedBaseline }, { AV_PROFILE_UNKNOWN } }; static const VAAPIEncodeType vaapi_encode_type_h264 = { .profiles = vaapi_encode_h264_profiles, .flags = FLAG_SLICE_CONTROL | FLAG_B_PICTURES | FLAG_B_PICTURE_REFERENCES | FLAG_NON_IDR_KEY_PICTURES, .default_quality = 20, .configure = &vaapi_encode_h264_configure, .picture_priv_data_size = sizeof(VAAPIEncodeH264Picture), .sequence_params_size = sizeof(VAEncSequenceParameterBufferH264), .init_sequence_params = &vaapi_encode_h264_init_sequence_params, .picture_params_size = sizeof(VAEncPictureParameterBufferH264), .init_picture_params = &vaapi_encode_h264_init_picture_params, .slice_params_size = sizeof(VAEncSliceParameterBufferH264), .init_slice_params = &vaapi_encode_h264_init_slice_params, .sequence_header_type = VAEncPackedHeaderSequence, .write_sequence_header = &vaapi_encode_h264_write_sequence_header, .slice_header_type = VAEncPackedHeaderH264_Slice, .write_slice_header = &vaapi_encode_h264_write_slice_header, .write_extra_header = &vaapi_encode_h264_write_extra_header, }; static av_cold int vaapi_encode_h264_init(AVCodecContext *avctx) { VAAPIEncodeContext *ctx = avctx->priv_data; VAAPIEncodeH264Context *priv = avctx->priv_data; ctx->codec = &vaapi_encode_type_h264; if (avctx->profile == AV_PROFILE_UNKNOWN) avctx->profile = priv->profile; if (avctx->level == AV_LEVEL_UNKNOWN) avctx->level = priv->level; if (avctx->compression_level == FF_COMPRESSION_DEFAULT) avctx->compression_level = priv->quality; // Reject unsupported profiles. switch (avctx->profile) { case AV_PROFILE_H264_BASELINE: av_log(avctx, AV_LOG_WARNING, "H.264 baseline profile is not " "supported, using constrained baseline profile instead.\n"); avctx->profile = AV_PROFILE_H264_CONSTRAINED_BASELINE; break; case AV_PROFILE_H264_EXTENDED: av_log(avctx, AV_LOG_ERROR, "H.264 extended profile " "is not supported.\n"); return AVERROR_PATCHWELCOME; case AV_PROFILE_H264_HIGH_10_INTRA: av_log(avctx, AV_LOG_ERROR, "H.264 high 10 intra profile " "is not supported.\n"); return AVERROR_PATCHWELCOME; case AV_PROFILE_H264_HIGH_422: case AV_PROFILE_H264_HIGH_422_INTRA: case AV_PROFILE_H264_HIGH_444: case AV_PROFILE_H264_HIGH_444_PREDICTIVE: case AV_PROFILE_H264_HIGH_444_INTRA: case AV_PROFILE_H264_CAVLC_444: av_log(avctx, AV_LOG_ERROR, "H.264 non-4:2:0 profiles " "are not supported.\n"); return AVERROR_PATCHWELCOME; } if (avctx->level != AV_LEVEL_UNKNOWN && avctx->level & ~0xff) { av_log(avctx, AV_LOG_ERROR, "Invalid level %d: must fit " "in 8-bit unsigned integer.\n", avctx->level); return AVERROR(EINVAL); } ctx->desired_packed_headers = VA_ENC_PACKED_HEADER_SEQUENCE | // SPS and PPS. VA_ENC_PACKED_HEADER_SLICE | // Slice headers. VA_ENC_PACKED_HEADER_MISC; // SEI. ctx->surface_width = FFALIGN(avctx->width, 16); ctx->surface_height = FFALIGN(avctx->height, 16); ctx->slice_block_height = ctx->slice_block_width = 16; if (priv->qp > 0) ctx->explicit_qp = priv->qp; return ff_vaapi_encode_init(avctx); } static av_cold int vaapi_encode_h264_close(AVCodecContext *avctx) { VAAPIEncodeH264Context *priv = avctx->priv_data; ff_cbs_fragment_free(&priv->current_access_unit); ff_cbs_close(&priv->cbc); av_freep(&priv->sei_identifier_string); av_freep(&priv->sei_a53cc_data); return ff_vaapi_encode_close(avctx); } #define OFFSET(x) offsetof(VAAPIEncodeH264Context, x) #define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM) static const AVOption vaapi_encode_h264_options[] = { VAAPI_ENCODE_COMMON_OPTIONS, VAAPI_ENCODE_RC_OPTIONS, { "qp", "Constant QP (for P-frames; scaled by qfactor/qoffset for I/B)", OFFSET(qp), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 52, FLAGS }, { "quality", "Set encode quality (trades off against speed, higher is faster)", OFFSET(quality), AV_OPT_TYPE_INT, { .i64 = -1 }, -1, INT_MAX, FLAGS }, { "coder", "Entropy coder type", OFFSET(coder), AV_OPT_TYPE_INT, { .i64 = 1 }, 0, 1, FLAGS, .unit = "coder" }, { "cavlc", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 0 }, INT_MIN, INT_MAX, FLAGS, .unit = "coder" }, { "cabac", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 1 }, INT_MIN, INT_MAX, FLAGS, .unit = "coder" }, { "vlc", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 0 }, INT_MIN, INT_MAX, FLAGS, .unit = "coder" }, { "ac", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = 1 }, INT_MIN, INT_MAX, FLAGS, .unit = "coder" }, { "aud", "Include AUD", OFFSET(aud), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, { "sei", "Set SEI to include", OFFSET(sei), AV_OPT_TYPE_FLAGS, { .i64 = SEI_IDENTIFIER | SEI_TIMING | SEI_RECOVERY_POINT | SEI_A53_CC }, 0, INT_MAX, FLAGS, .unit = "sei" }, { "identifier", "Include encoder version identifier", 0, AV_OPT_TYPE_CONST, { .i64 = SEI_IDENTIFIER }, INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, { "timing", "Include timing parameters (buffering_period and pic_timing)", 0, AV_OPT_TYPE_CONST, { .i64 = SEI_TIMING }, INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, { "recovery_point", "Include recovery points where appropriate", 0, AV_OPT_TYPE_CONST, { .i64 = SEI_RECOVERY_POINT }, INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, { "a53_cc", "Include A/53 caption data", 0, AV_OPT_TYPE_CONST, { .i64 = SEI_A53_CC }, INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, { "profile", "Set profile (profile_idc and constraint_set*_flag)", OFFSET(profile), AV_OPT_TYPE_INT, { .i64 = AV_PROFILE_UNKNOWN }, AV_PROFILE_UNKNOWN, 0xffff, FLAGS, .unit = "profile" }, #define PROFILE(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \ { .i64 = value }, 0, 0, FLAGS, .unit = "profile" { PROFILE("constrained_baseline", AV_PROFILE_H264_CONSTRAINED_BASELINE) }, { PROFILE("main", AV_PROFILE_H264_MAIN) }, { PROFILE("high", AV_PROFILE_H264_HIGH) }, { PROFILE("high10", AV_PROFILE_H264_HIGH_10) }, #undef PROFILE { "level", "Set level (level_idc)", OFFSET(level), AV_OPT_TYPE_INT, { .i64 = AV_LEVEL_UNKNOWN }, AV_LEVEL_UNKNOWN, 0xff, FLAGS, .unit = "level" }, #define LEVEL(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \ { .i64 = value }, 0, 0, FLAGS, .unit = "level" { LEVEL("1", 10) }, { LEVEL("1.1", 11) }, { LEVEL("1.2", 12) }, { LEVEL("1.3", 13) }, { LEVEL("2", 20) }, { LEVEL("2.1", 21) }, { LEVEL("2.2", 22) }, { LEVEL("3", 30) }, { LEVEL("3.1", 31) }, { LEVEL("3.2", 32) }, { LEVEL("4", 40) }, { LEVEL("4.1", 41) }, { LEVEL("4.2", 42) }, { LEVEL("5", 50) }, { LEVEL("5.1", 51) }, { LEVEL("5.2", 52) }, { LEVEL("6", 60) }, { LEVEL("6.1", 61) }, { LEVEL("6.2", 62) }, #undef LEVEL { NULL }, }; static const FFCodecDefault vaapi_encode_h264_defaults[] = { { "b", "0" }, { "bf", "2" }, { "g", "120" }, { "i_qfactor", "1" }, { "i_qoffset", "0" }, { "b_qfactor", "6/5" }, { "b_qoffset", "0" }, { "qmin", "-1" }, { "qmax", "-1" }, { NULL }, }; static const AVClass vaapi_encode_h264_class = { .class_name = "h264_vaapi", .item_name = av_default_item_name, .option = vaapi_encode_h264_options, .version = LIBAVUTIL_VERSION_INT, }; const FFCodec ff_h264_vaapi_encoder = { .p.name = "h264_vaapi", CODEC_LONG_NAME("H.264/AVC (VAAPI)"), .p.type = AVMEDIA_TYPE_VIDEO, .p.id = AV_CODEC_ID_H264, .priv_data_size = sizeof(VAAPIEncodeH264Context), .init = &vaapi_encode_h264_init, FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet), .close = &vaapi_encode_h264_close, .p.priv_class = &vaapi_encode_h264_class, .p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE | AV_CODEC_CAP_DR1 | AV_CODEC_CAP_ENCODER_REORDERED_OPAQUE, .caps_internal = FF_CODEC_CAP_NOT_INIT_THREADSAFE | FF_CODEC_CAP_INIT_CLEANUP, .defaults = vaapi_encode_h264_defaults, .p.pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_VAAPI, AV_PIX_FMT_NONE, }, .hw_configs = ff_vaapi_encode_hw_configs, .p.wrapper_name = "vaapi", };