mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1557 lines
55 KiB
1557 lines
55 KiB
/* |
|
* Direct3D 12 HW acceleration video encoder |
|
* |
|
* Copyright (c) 2024 Intel Corporation |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include "libavutil/avassert.h" |
|
#include "libavutil/common.h" |
|
#include "libavutil/internal.h" |
|
#include "libavutil/log.h" |
|
#include "libavutil/mem.h" |
|
#include "libavutil/pixdesc.h" |
|
#include "libavutil/hwcontext_d3d12va_internal.h" |
|
#include "libavutil/hwcontext_d3d12va.h" |
|
|
|
#include "avcodec.h" |
|
#include "d3d12va_encode.h" |
|
#include "encode.h" |
|
|
|
const AVCodecHWConfigInternal *const ff_d3d12va_encode_hw_configs[] = { |
|
HW_CONFIG_ENCODER_FRAMES(D3D12, D3D12VA), |
|
NULL, |
|
}; |
|
|
|
static int d3d12va_fence_completion(AVD3D12VASyncContext *psync_ctx) |
|
{ |
|
uint64_t completion = ID3D12Fence_GetCompletedValue(psync_ctx->fence); |
|
if (completion < psync_ctx->fence_value) { |
|
if (FAILED(ID3D12Fence_SetEventOnCompletion(psync_ctx->fence, psync_ctx->fence_value, psync_ctx->event))) |
|
return AVERROR(EINVAL); |
|
|
|
WaitForSingleObjectEx(psync_ctx->event, INFINITE, FALSE); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_sync_with_gpu(AVCodecContext *avctx) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
|
|
DX_CHECK(ID3D12CommandQueue_Signal(ctx->command_queue, ctx->sync_ctx.fence, ++ctx->sync_ctx.fence_value)); |
|
return d3d12va_fence_completion(&ctx->sync_ctx); |
|
|
|
fail: |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
typedef struct CommandAllocator { |
|
ID3D12CommandAllocator *command_allocator; |
|
uint64_t fence_value; |
|
} CommandAllocator; |
|
|
|
static int d3d12va_get_valid_command_allocator(AVCodecContext *avctx, ID3D12CommandAllocator **ppAllocator) |
|
{ |
|
HRESULT hr; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
CommandAllocator allocator; |
|
|
|
if (av_fifo_peek(ctx->allocator_queue, &allocator, 1, 0) >= 0) { |
|
uint64_t completion = ID3D12Fence_GetCompletedValue(ctx->sync_ctx.fence); |
|
if (completion >= allocator.fence_value) { |
|
*ppAllocator = allocator.command_allocator; |
|
av_fifo_read(ctx->allocator_queue, &allocator, 1); |
|
return 0; |
|
} |
|
} |
|
|
|
hr = ID3D12Device_CreateCommandAllocator(ctx->hwctx->device, D3D12_COMMAND_LIST_TYPE_VIDEO_ENCODE, |
|
&IID_ID3D12CommandAllocator, (void **)ppAllocator); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create a new command allocator!\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_discard_command_allocator(AVCodecContext *avctx, ID3D12CommandAllocator *pAllocator, uint64_t fence_value) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
|
|
CommandAllocator allocator = { |
|
.command_allocator = pAllocator, |
|
.fence_value = fence_value, |
|
}; |
|
|
|
av_fifo_write(ctx->allocator_queue, &allocator, 1); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_wait(AVCodecContext *avctx, |
|
FFHWBaseEncodePicture *base_pic) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12VAEncodePicture *pic = base_pic->priv; |
|
uint64_t completion; |
|
|
|
av_assert0(base_pic->encode_issued); |
|
|
|
if (base_pic->encode_complete) { |
|
// Already waited for this picture. |
|
return 0; |
|
} |
|
|
|
completion = ID3D12Fence_GetCompletedValue(ctx->sync_ctx.fence); |
|
if (completion < pic->fence_value) { |
|
if (FAILED(ID3D12Fence_SetEventOnCompletion(ctx->sync_ctx.fence, pic->fence_value, |
|
ctx->sync_ctx.event))) |
|
return AVERROR(EINVAL); |
|
|
|
WaitForSingleObjectEx(ctx->sync_ctx.event, INFINITE, FALSE); |
|
} |
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Sync to pic %"PRId64"/%"PRId64" " |
|
"(input surface %p).\n", base_pic->display_order, |
|
base_pic->encode_order, pic->input_surface->texture); |
|
|
|
av_frame_free(&base_pic->input_image); |
|
|
|
base_pic->encode_complete = 1; |
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_create_metadata_buffers(AVCodecContext *avctx, |
|
D3D12VAEncodePicture *pic) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
int width = sizeof(D3D12_VIDEO_ENCODER_OUTPUT_METADATA) + sizeof(D3D12_VIDEO_ENCODER_FRAME_SUBREGION_METADATA); |
|
D3D12_HEAP_PROPERTIES encoded_meta_props = { .Type = D3D12_HEAP_TYPE_DEFAULT }, resolved_meta_props; |
|
D3D12_HEAP_TYPE resolved_heap_type = D3D12_HEAP_TYPE_READBACK; |
|
HRESULT hr; |
|
|
|
D3D12_RESOURCE_DESC meta_desc = { |
|
.Dimension = D3D12_RESOURCE_DIMENSION_BUFFER, |
|
.Alignment = 0, |
|
.Width = ctx->req.MaxEncoderOutputMetadataBufferSize, |
|
.Height = 1, |
|
.DepthOrArraySize = 1, |
|
.MipLevels = 1, |
|
.Format = DXGI_FORMAT_UNKNOWN, |
|
.SampleDesc = { .Count = 1, .Quality = 0 }, |
|
.Layout = D3D12_TEXTURE_LAYOUT_ROW_MAJOR, |
|
.Flags = D3D12_RESOURCE_FLAG_NONE, |
|
}; |
|
|
|
hr = ID3D12Device_CreateCommittedResource(ctx->hwctx->device, &encoded_meta_props, D3D12_HEAP_FLAG_NONE, |
|
&meta_desc, D3D12_RESOURCE_STATE_COMMON, NULL, |
|
&IID_ID3D12Resource, (void **)&pic->encoded_metadata); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create metadata buffer.\n"); |
|
return AVERROR_UNKNOWN; |
|
} |
|
|
|
ctx->hwctx->device->lpVtbl->GetCustomHeapProperties(ctx->hwctx->device, &resolved_meta_props, 0, resolved_heap_type); |
|
|
|
meta_desc.Width = width; |
|
|
|
hr = ID3D12Device_CreateCommittedResource(ctx->hwctx->device, &resolved_meta_props, D3D12_HEAP_FLAG_NONE, |
|
&meta_desc, D3D12_RESOURCE_STATE_COMMON, NULL, |
|
&IID_ID3D12Resource, (void **)&pic->resolved_metadata); |
|
|
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create output metadata buffer.\n"); |
|
return AVERROR_UNKNOWN; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_issue(AVCodecContext *avctx, |
|
FFHWBaseEncodePicture *base_pic) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12VAEncodePicture *pic = base_pic->priv; |
|
AVD3D12VAFramesContext *frames_hwctx = base_ctx->input_frames->hwctx; |
|
int err, i, j; |
|
HRESULT hr; |
|
char data[MAX_PARAM_BUFFER_SIZE]; |
|
void *ptr; |
|
size_t bit_len; |
|
ID3D12CommandAllocator *command_allocator = NULL; |
|
ID3D12VideoEncodeCommandList2 *cmd_list = ctx->command_list; |
|
D3D12_RESOURCE_BARRIER barriers[32] = { 0 }; |
|
D3D12_VIDEO_ENCODE_REFERENCE_FRAMES d3d12_refs = { 0 }; |
|
|
|
D3D12_VIDEO_ENCODER_ENCODEFRAME_INPUT_ARGUMENTS input_args = { |
|
.SequenceControlDesc = { |
|
.Flags = D3D12_VIDEO_ENCODER_SEQUENCE_CONTROL_FLAG_NONE, |
|
.IntraRefreshConfig = { 0 }, |
|
.RateControl = ctx->rc, |
|
.PictureTargetResolution = ctx->resolution, |
|
.SelectedLayoutMode = D3D12_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE_FULL_FRAME, |
|
.FrameSubregionsLayoutData = { 0 }, |
|
.CodecGopSequence = ctx->gop, |
|
}, |
|
.pInputFrame = pic->input_surface->texture, |
|
.InputFrameSubresource = 0, |
|
}; |
|
|
|
D3D12_VIDEO_ENCODER_ENCODEFRAME_OUTPUT_ARGUMENTS output_args = { 0 }; |
|
|
|
D3D12_VIDEO_ENCODER_RESOLVE_METADATA_INPUT_ARGUMENTS input_metadata = { |
|
.EncoderCodec = ctx->codec->d3d12_codec, |
|
.EncoderProfile = ctx->profile->d3d12_profile, |
|
.EncoderInputFormat = frames_hwctx->format, |
|
.EncodedPictureEffectiveResolution = ctx->resolution, |
|
}; |
|
|
|
D3D12_VIDEO_ENCODER_RESOLVE_METADATA_OUTPUT_ARGUMENTS output_metadata = { 0 }; |
|
|
|
memset(data, 0, sizeof(data)); |
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Issuing encode for pic %"PRId64"/%"PRId64" " |
|
"as type %s.\n", base_pic->display_order, base_pic->encode_order, |
|
ff_hw_base_encode_get_pictype_name(base_pic->type)); |
|
if (base_pic->nb_refs[0] == 0 && base_pic->nb_refs[1] == 0) { |
|
av_log(avctx, AV_LOG_DEBUG, "No reference pictures.\n"); |
|
} else { |
|
av_log(avctx, AV_LOG_DEBUG, "L0 refers to"); |
|
for (i = 0; i < base_pic->nb_refs[0]; i++) { |
|
av_log(avctx, AV_LOG_DEBUG, " %"PRId64"/%"PRId64, |
|
base_pic->refs[0][i]->display_order, base_pic->refs[0][i]->encode_order); |
|
} |
|
av_log(avctx, AV_LOG_DEBUG, ".\n"); |
|
|
|
if (base_pic->nb_refs[1]) { |
|
av_log(avctx, AV_LOG_DEBUG, "L1 refers to"); |
|
for (i = 0; i < base_pic->nb_refs[1]; i++) { |
|
av_log(avctx, AV_LOG_DEBUG, " %"PRId64"/%"PRId64, |
|
base_pic->refs[1][i]->display_order, base_pic->refs[1][i]->encode_order); |
|
} |
|
av_log(avctx, AV_LOG_DEBUG, ".\n"); |
|
} |
|
} |
|
|
|
av_assert0(!base_pic->encode_issued); |
|
for (i = 0; i < base_pic->nb_refs[0]; i++) { |
|
av_assert0(base_pic->refs[0][i]); |
|
av_assert0(base_pic->refs[0][i]->encode_issued); |
|
} |
|
for (i = 0; i < base_pic->nb_refs[1]; i++) { |
|
av_assert0(base_pic->refs[1][i]); |
|
av_assert0(base_pic->refs[1][i]->encode_issued); |
|
} |
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Input surface is %p.\n", pic->input_surface->texture); |
|
|
|
pic->recon_surface = (AVD3D12VAFrame *)base_pic->recon_image->data[0]; |
|
av_log(avctx, AV_LOG_DEBUG, "Recon surface is %p.\n", |
|
pic->recon_surface->texture); |
|
|
|
pic->output_buffer_ref = av_buffer_pool_get(ctx->output_buffer_pool); |
|
if (!pic->output_buffer_ref) { |
|
err = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
pic->output_buffer = (ID3D12Resource *)pic->output_buffer_ref->data; |
|
av_log(avctx, AV_LOG_DEBUG, "Output buffer is %p.\n", |
|
pic->output_buffer); |
|
|
|
err = d3d12va_encode_create_metadata_buffers(avctx, pic); |
|
if (err < 0) |
|
goto fail; |
|
|
|
if (ctx->codec->init_picture_params) { |
|
err = ctx->codec->init_picture_params(avctx, base_pic); |
|
if (err < 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to initialise picture " |
|
"parameters: %d.\n", err); |
|
goto fail; |
|
} |
|
} |
|
|
|
if (base_pic->type == FF_HW_PICTURE_TYPE_IDR) { |
|
if (ctx->codec->write_sequence_header) { |
|
bit_len = 8 * sizeof(data); |
|
err = ctx->codec->write_sequence_header(avctx, data, &bit_len); |
|
if (err < 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to write per-sequence " |
|
"header: %d.\n", err); |
|
goto fail; |
|
} |
|
} |
|
|
|
pic->header_size = (int)bit_len / 8; |
|
pic->aligned_header_size = pic->header_size % ctx->req.CompressedBitstreamBufferAccessAlignment ? |
|
FFALIGN(pic->header_size, ctx->req.CompressedBitstreamBufferAccessAlignment) : |
|
pic->header_size; |
|
|
|
hr = ID3D12Resource_Map(pic->output_buffer, 0, NULL, (void **)&ptr); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
memcpy(ptr, data, pic->aligned_header_size); |
|
ID3D12Resource_Unmap(pic->output_buffer, 0, NULL); |
|
} |
|
|
|
d3d12_refs.NumTexture2Ds = base_pic->nb_refs[0] + base_pic->nb_refs[1]; |
|
if (d3d12_refs.NumTexture2Ds) { |
|
d3d12_refs.ppTexture2Ds = av_calloc(d3d12_refs.NumTexture2Ds, |
|
sizeof(*d3d12_refs.ppTexture2Ds)); |
|
if (!d3d12_refs.ppTexture2Ds) { |
|
err = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
i = 0; |
|
for (j = 0; j < base_pic->nb_refs[0]; j++) |
|
d3d12_refs.ppTexture2Ds[i++] = ((D3D12VAEncodePicture *)base_pic->refs[0][j]->priv)->recon_surface->texture; |
|
for (j = 0; j < base_pic->nb_refs[1]; j++) |
|
d3d12_refs.ppTexture2Ds[i++] = ((D3D12VAEncodePicture *)base_pic->refs[1][j]->priv)->recon_surface->texture; |
|
} |
|
|
|
input_args.PictureControlDesc.IntraRefreshFrameIndex = 0; |
|
if (base_pic->is_reference) |
|
input_args.PictureControlDesc.Flags |= D3D12_VIDEO_ENCODER_PICTURE_CONTROL_FLAG_USED_AS_REFERENCE_PICTURE; |
|
|
|
input_args.PictureControlDesc.PictureControlCodecData = pic->pic_ctl; |
|
input_args.PictureControlDesc.ReferenceFrames = d3d12_refs; |
|
input_args.CurrentFrameBitstreamMetadataSize = pic->aligned_header_size; |
|
|
|
output_args.Bitstream.pBuffer = pic->output_buffer; |
|
output_args.Bitstream.FrameStartOffset = pic->aligned_header_size; |
|
output_args.ReconstructedPicture.pReconstructedPicture = pic->recon_surface->texture; |
|
output_args.ReconstructedPicture.ReconstructedPictureSubresource = 0; |
|
output_args.EncoderOutputMetadata.pBuffer = pic->encoded_metadata; |
|
output_args.EncoderOutputMetadata.Offset = 0; |
|
|
|
input_metadata.HWLayoutMetadata.pBuffer = pic->encoded_metadata; |
|
input_metadata.HWLayoutMetadata.Offset = 0; |
|
|
|
output_metadata.ResolvedLayoutMetadata.pBuffer = pic->resolved_metadata; |
|
output_metadata.ResolvedLayoutMetadata.Offset = 0; |
|
|
|
err = d3d12va_get_valid_command_allocator(avctx, &command_allocator); |
|
if (err < 0) |
|
goto fail; |
|
|
|
hr = ID3D12CommandAllocator_Reset(command_allocator); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12VideoEncodeCommandList2_Reset(cmd_list, command_allocator); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
#define TRANSITION_BARRIER(res, before, after) \ |
|
(D3D12_RESOURCE_BARRIER) { \ |
|
.Type = D3D12_RESOURCE_BARRIER_TYPE_TRANSITION, \ |
|
.Flags = D3D12_RESOURCE_BARRIER_FLAG_NONE, \ |
|
.Transition = { \ |
|
.pResource = res, \ |
|
.Subresource = D3D12_RESOURCE_BARRIER_ALL_SUBRESOURCES, \ |
|
.StateBefore = before, \ |
|
.StateAfter = after, \ |
|
}, \ |
|
} |
|
|
|
barriers[0] = TRANSITION_BARRIER(pic->input_surface->texture, |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ); |
|
barriers[1] = TRANSITION_BARRIER(pic->output_buffer, |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE); |
|
barriers[2] = TRANSITION_BARRIER(pic->recon_surface->texture, |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE); |
|
barriers[3] = TRANSITION_BARRIER(pic->encoded_metadata, |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE); |
|
barriers[4] = TRANSITION_BARRIER(pic->resolved_metadata, |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE); |
|
|
|
ID3D12VideoEncodeCommandList2_ResourceBarrier(cmd_list, 5, barriers); |
|
|
|
if (d3d12_refs.NumTexture2Ds) { |
|
D3D12_RESOURCE_BARRIER refs_barriers[3]; |
|
|
|
for (i = 0; i < d3d12_refs.NumTexture2Ds; i++) |
|
refs_barriers[i] = TRANSITION_BARRIER(d3d12_refs.ppTexture2Ds[i], |
|
D3D12_RESOURCE_STATE_COMMON, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ); |
|
|
|
ID3D12VideoEncodeCommandList2_ResourceBarrier(cmd_list, d3d12_refs.NumTexture2Ds, |
|
refs_barriers); |
|
} |
|
|
|
ID3D12VideoEncodeCommandList2_EncodeFrame(cmd_list, ctx->encoder, ctx->encoder_heap, |
|
&input_args, &output_args); |
|
|
|
barriers[3] = TRANSITION_BARRIER(pic->encoded_metadata, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ); |
|
|
|
ID3D12VideoEncodeCommandList2_ResourceBarrier(cmd_list, 1, &barriers[3]); |
|
|
|
ID3D12VideoEncodeCommandList2_ResolveEncoderOutputMetadata(cmd_list, &input_metadata, &output_metadata); |
|
|
|
if (d3d12_refs.NumTexture2Ds) { |
|
D3D12_RESOURCE_BARRIER refs_barriers[3]; |
|
|
|
for (i = 0; i < d3d12_refs.NumTexture2Ds; i++) |
|
refs_barriers[i] = TRANSITION_BARRIER(d3d12_refs.ppTexture2Ds[i], |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
|
|
ID3D12VideoEncodeCommandList2_ResourceBarrier(cmd_list, d3d12_refs.NumTexture2Ds, |
|
refs_barriers); |
|
} |
|
|
|
barriers[0] = TRANSITION_BARRIER(pic->input_surface->texture, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
barriers[1] = TRANSITION_BARRIER(pic->output_buffer, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
barriers[2] = TRANSITION_BARRIER(pic->recon_surface->texture, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
barriers[3] = TRANSITION_BARRIER(pic->encoded_metadata, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_READ, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
barriers[4] = TRANSITION_BARRIER(pic->resolved_metadata, |
|
D3D12_RESOURCE_STATE_VIDEO_ENCODE_WRITE, |
|
D3D12_RESOURCE_STATE_COMMON); |
|
|
|
ID3D12VideoEncodeCommandList2_ResourceBarrier(cmd_list, 5, barriers); |
|
|
|
hr = ID3D12VideoEncodeCommandList2_Close(cmd_list); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12CommandQueue_Wait(ctx->command_queue, pic->input_surface->sync_ctx.fence, |
|
pic->input_surface->sync_ctx.fence_value); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
ID3D12CommandQueue_ExecuteCommandLists(ctx->command_queue, 1, (ID3D12CommandList **)&ctx->command_list); |
|
|
|
hr = ID3D12CommandQueue_Signal(ctx->command_queue, pic->input_surface->sync_ctx.fence, |
|
++pic->input_surface->sync_ctx.fence_value); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12CommandQueue_Signal(ctx->command_queue, ctx->sync_ctx.fence, ++ctx->sync_ctx.fence_value); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
err = d3d12va_discard_command_allocator(avctx, command_allocator, ctx->sync_ctx.fence_value); |
|
if (err < 0) |
|
goto fail; |
|
|
|
pic->fence_value = ctx->sync_ctx.fence_value; |
|
|
|
if (d3d12_refs.ppTexture2Ds) |
|
av_freep(&d3d12_refs.ppTexture2Ds); |
|
|
|
return 0; |
|
|
|
fail: |
|
if (command_allocator) |
|
d3d12va_discard_command_allocator(avctx, command_allocator, ctx->sync_ctx.fence_value); |
|
|
|
if (d3d12_refs.ppTexture2Ds) |
|
av_freep(&d3d12_refs.ppTexture2Ds); |
|
|
|
if (ctx->codec->free_picture_params) |
|
ctx->codec->free_picture_params(pic); |
|
|
|
av_buffer_unref(&pic->output_buffer_ref); |
|
pic->output_buffer = NULL; |
|
D3D12_OBJECT_RELEASE(pic->encoded_metadata); |
|
D3D12_OBJECT_RELEASE(pic->resolved_metadata); |
|
return err; |
|
} |
|
|
|
static int d3d12va_encode_discard(AVCodecContext *avctx, |
|
FFHWBaseEncodePicture *base_pic) |
|
{ |
|
D3D12VAEncodePicture *pic = base_pic->priv; |
|
|
|
d3d12va_encode_wait(avctx, base_pic); |
|
|
|
if (pic->output_buffer_ref) { |
|
av_log(avctx, AV_LOG_DEBUG, "Discard output for pic " |
|
"%"PRId64"/%"PRId64".\n", |
|
base_pic->display_order, base_pic->encode_order); |
|
|
|
av_buffer_unref(&pic->output_buffer_ref); |
|
pic->output_buffer = NULL; |
|
} |
|
|
|
D3D12_OBJECT_RELEASE(pic->encoded_metadata); |
|
D3D12_OBJECT_RELEASE(pic->resolved_metadata); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_free_rc_params(AVCodecContext *avctx) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
|
|
switch (ctx->rc.Mode) |
|
{ |
|
case D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_CQP: |
|
av_freep(&ctx->rc.ConfigParams.pConfiguration_CQP); |
|
break; |
|
case D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_CBR: |
|
av_freep(&ctx->rc.ConfigParams.pConfiguration_CBR); |
|
break; |
|
case D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_VBR: |
|
av_freep(&ctx->rc.ConfigParams.pConfiguration_VBR); |
|
break; |
|
case D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_QVBR: |
|
av_freep(&ctx->rc.ConfigParams.pConfiguration_QVBR); |
|
break; |
|
default: |
|
break; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_init(AVCodecContext *avctx, FFHWBaseEncodePicture *pic) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12VAEncodePicture *priv = pic->priv; |
|
AVFrame *frame = pic->input_image; |
|
|
|
if (ctx->codec->picture_priv_data_size > 0) { |
|
pic->codec_priv = av_mallocz(ctx->codec->picture_priv_data_size); |
|
if (!pic->codec_priv) |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
priv->input_surface = (AVD3D12VAFrame *)frame->data[0]; |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_free(AVCodecContext *avctx, FFHWBaseEncodePicture *pic) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12VAEncodePicture *priv = pic->priv; |
|
|
|
if (pic->encode_issued) |
|
d3d12va_encode_discard(avctx, pic); |
|
|
|
if (ctx->codec->free_picture_params) |
|
ctx->codec->free_picture_params(priv); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_get_buffer_size(AVCodecContext *avctx, |
|
D3D12VAEncodePicture *pic, size_t *size) |
|
{ |
|
D3D12_VIDEO_ENCODER_OUTPUT_METADATA *meta = NULL; |
|
uint8_t *data; |
|
HRESULT hr; |
|
int err; |
|
|
|
hr = ID3D12Resource_Map(pic->resolved_metadata, 0, NULL, (void **)&data); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
return err; |
|
} |
|
|
|
meta = (D3D12_VIDEO_ENCODER_OUTPUT_METADATA *)data; |
|
|
|
if (meta->EncodeErrorFlags != D3D12_VIDEO_ENCODER_ENCODE_ERROR_FLAG_NO_ERROR) { |
|
av_log(avctx, AV_LOG_ERROR, "Encode failed %"PRIu64"\n", meta->EncodeErrorFlags); |
|
err = AVERROR(EINVAL); |
|
return err; |
|
} |
|
|
|
if (meta->EncodedBitstreamWrittenBytesCount == 0) { |
|
av_log(avctx, AV_LOG_ERROR, "No bytes were written to encoded bitstream\n"); |
|
err = AVERROR(EINVAL); |
|
return err; |
|
} |
|
|
|
*size = meta->EncodedBitstreamWrittenBytesCount; |
|
|
|
ID3D12Resource_Unmap(pic->resolved_metadata, 0, NULL); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_get_coded_data(AVCodecContext *avctx, |
|
D3D12VAEncodePicture *pic, AVPacket *pkt) |
|
{ |
|
int err; |
|
uint8_t *ptr, *mapped_data; |
|
size_t total_size = 0; |
|
HRESULT hr; |
|
|
|
err = d3d12va_encode_get_buffer_size(avctx, pic, &total_size); |
|
if (err < 0) |
|
goto end; |
|
|
|
total_size += pic->header_size; |
|
av_log(avctx, AV_LOG_DEBUG, "Output buffer size %"PRId64"\n", total_size); |
|
|
|
hr = ID3D12Resource_Map(pic->output_buffer, 0, NULL, (void **)&mapped_data); |
|
if (FAILED(hr)) { |
|
err = AVERROR_UNKNOWN; |
|
goto end; |
|
} |
|
|
|
err = ff_get_encode_buffer(avctx, pkt, total_size, 0); |
|
if (err < 0) |
|
goto end; |
|
ptr = pkt->data; |
|
|
|
memcpy(ptr, mapped_data, pic->header_size); |
|
|
|
ptr += pic->header_size; |
|
mapped_data += pic->aligned_header_size; |
|
total_size -= pic->header_size; |
|
|
|
memcpy(ptr, mapped_data, total_size); |
|
|
|
ID3D12Resource_Unmap(pic->output_buffer, 0, NULL); |
|
|
|
end: |
|
av_buffer_unref(&pic->output_buffer_ref); |
|
pic->output_buffer = NULL; |
|
return err; |
|
} |
|
|
|
static int d3d12va_encode_output(AVCodecContext *avctx, |
|
FFHWBaseEncodePicture *base_pic, AVPacket *pkt) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodePicture *pic = base_pic->priv; |
|
AVPacket *pkt_ptr = pkt; |
|
int err; |
|
|
|
err = d3d12va_encode_wait(avctx, base_pic); |
|
if (err < 0) |
|
return err; |
|
|
|
err = d3d12va_encode_get_coded_data(avctx, pic, pkt); |
|
if (err < 0) |
|
return err; |
|
|
|
av_log(avctx, AV_LOG_DEBUG, "Output read for pic %"PRId64"/%"PRId64".\n", |
|
base_pic->display_order, base_pic->encode_order); |
|
|
|
ff_hw_base_encode_set_output_property(base_ctx, avctx, (FFHWBaseEncodePicture *)base_pic, |
|
pkt_ptr, 0); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_set_profile(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
const D3D12VAEncodeProfile *profile; |
|
const AVPixFmtDescriptor *desc; |
|
int i, depth; |
|
|
|
desc = av_pix_fmt_desc_get(base_ctx->input_frames->sw_format); |
|
if (!desc) { |
|
av_log(avctx, AV_LOG_ERROR, "Invalid input pixfmt (%d).\n", |
|
base_ctx->input_frames->sw_format); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
depth = desc->comp[0].depth; |
|
for (i = 1; i < desc->nb_components; i++) { |
|
if (desc->comp[i].depth != depth) { |
|
av_log(avctx, AV_LOG_ERROR, "Invalid input pixfmt (%s).\n", |
|
desc->name); |
|
return AVERROR(EINVAL); |
|
} |
|
} |
|
av_log(avctx, AV_LOG_VERBOSE, "Input surface format is %s.\n", |
|
desc->name); |
|
|
|
av_assert0(ctx->codec->profiles); |
|
for (i = 0; (ctx->codec->profiles[i].av_profile != |
|
AV_PROFILE_UNKNOWN); i++) { |
|
profile = &ctx->codec->profiles[i]; |
|
if (depth != profile->depth || |
|
desc->nb_components != profile->nb_components) |
|
continue; |
|
if (desc->nb_components > 1 && |
|
(desc->log2_chroma_w != profile->log2_chroma_w || |
|
desc->log2_chroma_h != profile->log2_chroma_h)) |
|
continue; |
|
if (avctx->profile != profile->av_profile && |
|
avctx->profile != AV_PROFILE_UNKNOWN) |
|
continue; |
|
|
|
ctx->profile = profile; |
|
break; |
|
} |
|
if (!ctx->profile) { |
|
av_log(avctx, AV_LOG_ERROR, "No usable encoding profile found.\n"); |
|
return AVERROR(ENOSYS); |
|
} |
|
|
|
avctx->profile = profile->av_profile; |
|
return 0; |
|
} |
|
|
|
static const D3D12VAEncodeRCMode d3d12va_encode_rc_modes[] = { |
|
// Bitrate Quality |
|
// | Maxrate | HRD/VBV |
|
{ 0 }, // | | | | |
|
{ RC_MODE_CQP, "CQP", 0, 0, 1, 0, D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_CQP }, |
|
{ RC_MODE_CBR, "CBR", 1, 0, 0, 1, D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_CBR }, |
|
{ RC_MODE_VBR, "VBR", 1, 1, 0, 1, D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_VBR }, |
|
{ RC_MODE_QVBR, "QVBR", 1, 1, 1, 1, D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_QVBR }, |
|
}; |
|
|
|
static int check_rate_control_support(AVCodecContext *avctx, const D3D12VAEncodeRCMode *rc_mode) |
|
{ |
|
HRESULT hr; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12_FEATURE_DATA_VIDEO_ENCODER_RATE_CONTROL_MODE d3d12_rc_mode = { |
|
.Codec = ctx->codec->d3d12_codec, |
|
}; |
|
|
|
if (!rc_mode->d3d12_mode) |
|
return 0; |
|
|
|
d3d12_rc_mode.IsSupported = 0; |
|
d3d12_rc_mode.RateControlMode = rc_mode->d3d12_mode; |
|
|
|
hr = ID3D12VideoDevice3_CheckFeatureSupport(ctx->video_device3, |
|
D3D12_FEATURE_VIDEO_ENCODER_RATE_CONTROL_MODE, |
|
&d3d12_rc_mode, sizeof(d3d12_rc_mode)); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to check rate control support.\n"); |
|
return 0; |
|
} |
|
|
|
return d3d12_rc_mode.IsSupported; |
|
} |
|
|
|
static int d3d12va_encode_init_rate_control(AVCodecContext *avctx) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
int64_t rc_target_bitrate; |
|
int64_t rc_peak_bitrate; |
|
int rc_quality; |
|
int64_t hrd_buffer_size; |
|
int64_t hrd_initial_buffer_fullness; |
|
int fr_num, fr_den; |
|
const D3D12VAEncodeRCMode *rc_mode; |
|
|
|
// Rate control mode selection: |
|
// * If the user has set a mode explicitly with the rc_mode option, |
|
// use it and fail if it is not available. |
|
// * If an explicit QP option has been set, use CQP. |
|
// * If the codec is CQ-only, use CQP. |
|
// * If the QSCALE avcodec option is set, use CQP. |
|
// * If bitrate and quality are both set, try QVBR. |
|
// * If quality is set, try CQP. |
|
// * If bitrate and maxrate are set and have the same value, try CBR. |
|
// * If a bitrate is set, try VBR, then CBR. |
|
// * If no bitrate is set, try CQP. |
|
|
|
#define TRY_RC_MODE(mode, fail) do { \ |
|
rc_mode = &d3d12va_encode_rc_modes[mode]; \ |
|
if (!(rc_mode->d3d12_mode && check_rate_control_support(avctx, rc_mode))) { \ |
|
if (fail) { \ |
|
av_log(avctx, AV_LOG_ERROR, "Driver does not support %s " \ |
|
"RC mode.\n", rc_mode->name); \ |
|
return AVERROR(EINVAL); \ |
|
} \ |
|
av_log(avctx, AV_LOG_DEBUG, "Driver does not support %s " \ |
|
"RC mode.\n", rc_mode->name); \ |
|
rc_mode = NULL; \ |
|
} else { \ |
|
goto rc_mode_found; \ |
|
} \ |
|
} while (0) |
|
|
|
if (ctx->explicit_rc_mode) |
|
TRY_RC_MODE(ctx->explicit_rc_mode, 1); |
|
|
|
if (ctx->explicit_qp) |
|
TRY_RC_MODE(RC_MODE_CQP, 1); |
|
|
|
if (ctx->codec->flags & FF_HW_FLAG_CONSTANT_QUALITY_ONLY) |
|
TRY_RC_MODE(RC_MODE_CQP, 1); |
|
|
|
if (avctx->flags & AV_CODEC_FLAG_QSCALE) |
|
TRY_RC_MODE(RC_MODE_CQP, 1); |
|
|
|
if (avctx->bit_rate > 0 && avctx->global_quality > 0) |
|
TRY_RC_MODE(RC_MODE_QVBR, 0); |
|
|
|
if (avctx->global_quality > 0) { |
|
TRY_RC_MODE(RC_MODE_CQP, 0); |
|
} |
|
|
|
if (avctx->bit_rate > 0 && avctx->rc_max_rate == avctx->bit_rate) |
|
TRY_RC_MODE(RC_MODE_CBR, 0); |
|
|
|
if (avctx->bit_rate > 0) { |
|
TRY_RC_MODE(RC_MODE_VBR, 0); |
|
TRY_RC_MODE(RC_MODE_CBR, 0); |
|
} else { |
|
TRY_RC_MODE(RC_MODE_CQP, 0); |
|
} |
|
|
|
av_log(avctx, AV_LOG_ERROR, "Driver does not support any " |
|
"RC mode compatible with selected options.\n"); |
|
return AVERROR(EINVAL); |
|
|
|
rc_mode_found: |
|
if (rc_mode->bitrate) { |
|
if (avctx->bit_rate <= 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Bitrate must be set for %s " |
|
"RC mode.\n", rc_mode->name); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (rc_mode->maxrate) { |
|
if (avctx->rc_max_rate > 0) { |
|
if (avctx->rc_max_rate < avctx->bit_rate) { |
|
av_log(avctx, AV_LOG_ERROR, "Invalid bitrate settings: " |
|
"bitrate (%"PRId64") must not be greater than " |
|
"maxrate (%"PRId64").\n", avctx->bit_rate, |
|
avctx->rc_max_rate); |
|
return AVERROR(EINVAL); |
|
} |
|
rc_target_bitrate = avctx->bit_rate; |
|
rc_peak_bitrate = avctx->rc_max_rate; |
|
} else { |
|
// We only have a target bitrate, but this mode requires |
|
// that a maximum rate be supplied as well. Since the |
|
// user does not want this to be a constraint, arbitrarily |
|
// pick a maximum rate of double the target rate. |
|
rc_target_bitrate = avctx->bit_rate; |
|
rc_peak_bitrate = 2 * avctx->bit_rate; |
|
} |
|
} else { |
|
if (avctx->rc_max_rate > avctx->bit_rate) { |
|
av_log(avctx, AV_LOG_WARNING, "Max bitrate is ignored " |
|
"in %s RC mode.\n", rc_mode->name); |
|
} |
|
rc_target_bitrate = avctx->bit_rate; |
|
rc_peak_bitrate = 0; |
|
} |
|
} else { |
|
rc_target_bitrate = 0; |
|
rc_peak_bitrate = 0; |
|
} |
|
|
|
if (rc_mode->quality) { |
|
if (ctx->explicit_qp) { |
|
rc_quality = ctx->explicit_qp; |
|
} else if (avctx->global_quality > 0) { |
|
if (avctx->flags & AV_CODEC_FLAG_QSCALE) |
|
rc_quality = avctx->global_quality / FF_QP2LAMBDA; |
|
else |
|
rc_quality = avctx->global_quality; |
|
} else { |
|
rc_quality = ctx->codec->default_quality; |
|
av_log(avctx, AV_LOG_WARNING, "No quality level set; " |
|
"using default (%d).\n", rc_quality); |
|
} |
|
} else { |
|
rc_quality = 0; |
|
} |
|
|
|
if (rc_mode->hrd) { |
|
if (avctx->rc_buffer_size) |
|
hrd_buffer_size = avctx->rc_buffer_size; |
|
else if (avctx->rc_max_rate > 0) |
|
hrd_buffer_size = avctx->rc_max_rate; |
|
else |
|
hrd_buffer_size = avctx->bit_rate; |
|
if (avctx->rc_initial_buffer_occupancy) { |
|
if (avctx->rc_initial_buffer_occupancy > hrd_buffer_size) { |
|
av_log(avctx, AV_LOG_ERROR, "Invalid RC buffer settings: " |
|
"must have initial buffer size (%d) <= " |
|
"buffer size (%"PRId64").\n", |
|
avctx->rc_initial_buffer_occupancy, hrd_buffer_size); |
|
return AVERROR(EINVAL); |
|
} |
|
hrd_initial_buffer_fullness = avctx->rc_initial_buffer_occupancy; |
|
} else { |
|
hrd_initial_buffer_fullness = hrd_buffer_size * 3 / 4; |
|
} |
|
} else { |
|
if (avctx->rc_buffer_size || avctx->rc_initial_buffer_occupancy) { |
|
av_log(avctx, AV_LOG_WARNING, "Buffering settings are ignored " |
|
"in %s RC mode.\n", rc_mode->name); |
|
} |
|
|
|
hrd_buffer_size = 0; |
|
hrd_initial_buffer_fullness = 0; |
|
} |
|
|
|
if (rc_target_bitrate > UINT32_MAX || |
|
hrd_buffer_size > UINT32_MAX || |
|
hrd_initial_buffer_fullness > UINT32_MAX) { |
|
av_log(avctx, AV_LOG_ERROR, "RC parameters of 2^32 or " |
|
"greater are not supported by D3D12.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
ctx->rc_quality = rc_quality; |
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "RC mode: %s.\n", rc_mode->name); |
|
|
|
if (rc_mode->quality) |
|
av_log(avctx, AV_LOG_VERBOSE, "RC quality: %d.\n", rc_quality); |
|
|
|
if (rc_mode->hrd) { |
|
av_log(avctx, AV_LOG_VERBOSE, "RC buffer: %"PRId64" bits, " |
|
"initial fullness %"PRId64" bits.\n", |
|
hrd_buffer_size, hrd_initial_buffer_fullness); |
|
} |
|
|
|
if (avctx->framerate.num > 0 && avctx->framerate.den > 0) |
|
av_reduce(&fr_num, &fr_den, |
|
avctx->framerate.num, avctx->framerate.den, 65535); |
|
else |
|
av_reduce(&fr_num, &fr_den, |
|
avctx->time_base.den, avctx->time_base.num, 65535); |
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "RC framerate: %d/%d (%.2f fps).\n", |
|
fr_num, fr_den, (double)fr_num / fr_den); |
|
|
|
ctx->rc.Flags = D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_NONE; |
|
ctx->rc.TargetFrameRate.Numerator = fr_num; |
|
ctx->rc.TargetFrameRate.Denominator = fr_den; |
|
ctx->rc.Mode = rc_mode->d3d12_mode; |
|
|
|
switch (rc_mode->mode) { |
|
case RC_MODE_CQP: |
|
// cqp ConfigParams will be updated in ctx->codec->configure. |
|
break; |
|
|
|
case RC_MODE_CBR: |
|
D3D12_VIDEO_ENCODER_RATE_CONTROL_CBR *cbr_ctl; |
|
|
|
ctx->rc.ConfigParams.DataSize = sizeof(D3D12_VIDEO_ENCODER_RATE_CONTROL_CBR); |
|
cbr_ctl = av_mallocz(ctx->rc.ConfigParams.DataSize); |
|
if (!cbr_ctl) |
|
return AVERROR(ENOMEM); |
|
|
|
cbr_ctl->TargetBitRate = rc_target_bitrate; |
|
cbr_ctl->VBVCapacity = hrd_buffer_size; |
|
cbr_ctl->InitialVBVFullness = hrd_initial_buffer_fullness; |
|
ctx->rc.Flags |= D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_ENABLE_VBV_SIZES; |
|
|
|
if (avctx->qmin > 0 || avctx->qmax > 0) { |
|
cbr_ctl->MinQP = avctx->qmin; |
|
cbr_ctl->MaxQP = avctx->qmax; |
|
ctx->rc.Flags |= D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_ENABLE_QP_RANGE; |
|
} |
|
|
|
ctx->rc.ConfigParams.pConfiguration_CBR = cbr_ctl; |
|
break; |
|
|
|
case RC_MODE_VBR: |
|
D3D12_VIDEO_ENCODER_RATE_CONTROL_VBR *vbr_ctl; |
|
|
|
ctx->rc.ConfigParams.DataSize = sizeof(D3D12_VIDEO_ENCODER_RATE_CONTROL_VBR); |
|
vbr_ctl = av_mallocz(ctx->rc.ConfigParams.DataSize); |
|
if (!vbr_ctl) |
|
return AVERROR(ENOMEM); |
|
|
|
vbr_ctl->TargetAvgBitRate = rc_target_bitrate; |
|
vbr_ctl->PeakBitRate = rc_peak_bitrate; |
|
vbr_ctl->VBVCapacity = hrd_buffer_size; |
|
vbr_ctl->InitialVBVFullness = hrd_initial_buffer_fullness; |
|
ctx->rc.Flags |= D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_ENABLE_VBV_SIZES; |
|
|
|
if (avctx->qmin > 0 || avctx->qmax > 0) { |
|
vbr_ctl->MinQP = avctx->qmin; |
|
vbr_ctl->MaxQP = avctx->qmax; |
|
ctx->rc.Flags |= D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_ENABLE_QP_RANGE; |
|
} |
|
|
|
ctx->rc.ConfigParams.pConfiguration_VBR = vbr_ctl; |
|
break; |
|
|
|
case RC_MODE_QVBR: |
|
D3D12_VIDEO_ENCODER_RATE_CONTROL_QVBR *qvbr_ctl; |
|
|
|
ctx->rc.ConfigParams.DataSize = sizeof(D3D12_VIDEO_ENCODER_RATE_CONTROL_QVBR); |
|
qvbr_ctl = av_mallocz(ctx->rc.ConfigParams.DataSize); |
|
if (!qvbr_ctl) |
|
return AVERROR(ENOMEM); |
|
|
|
qvbr_ctl->TargetAvgBitRate = rc_target_bitrate; |
|
qvbr_ctl->PeakBitRate = rc_peak_bitrate; |
|
qvbr_ctl->ConstantQualityTarget = rc_quality; |
|
|
|
if (avctx->qmin > 0 || avctx->qmax > 0) { |
|
qvbr_ctl->MinQP = avctx->qmin; |
|
qvbr_ctl->MaxQP = avctx->qmax; |
|
ctx->rc.Flags |= D3D12_VIDEO_ENCODER_RATE_CONTROL_FLAG_ENABLE_QP_RANGE; |
|
} |
|
|
|
ctx->rc.ConfigParams.pConfiguration_QVBR = qvbr_ctl; |
|
break; |
|
|
|
default: |
|
break; |
|
} |
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_init_gop_structure(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
uint32_t ref_l0, ref_l1; |
|
int err; |
|
HRESULT hr; |
|
D3D12_FEATURE_DATA_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT support; |
|
union { |
|
D3D12_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT_H264 h264; |
|
D3D12_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT_HEVC hevc; |
|
} codec_support; |
|
|
|
support.NodeIndex = 0; |
|
support.Codec = ctx->codec->d3d12_codec; |
|
support.Profile = ctx->profile->d3d12_profile; |
|
|
|
switch (ctx->codec->d3d12_codec) { |
|
case D3D12_VIDEO_ENCODER_CODEC_H264: |
|
support.PictureSupport.DataSize = sizeof(codec_support.h264); |
|
support.PictureSupport.pH264Support = &codec_support.h264; |
|
break; |
|
|
|
case D3D12_VIDEO_ENCODER_CODEC_HEVC: |
|
support.PictureSupport.DataSize = sizeof(codec_support.hevc); |
|
support.PictureSupport.pHEVCSupport = &codec_support.hevc; |
|
break; |
|
|
|
default: |
|
av_assert0(0); |
|
} |
|
|
|
hr = ID3D12VideoDevice3_CheckFeatureSupport(ctx->video_device3, D3D12_FEATURE_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT, |
|
&support, sizeof(support)); |
|
if (FAILED(hr)) |
|
return AVERROR(EINVAL); |
|
|
|
if (support.IsSupported) { |
|
switch (ctx->codec->d3d12_codec) { |
|
case D3D12_VIDEO_ENCODER_CODEC_H264: |
|
ref_l0 = FFMIN(support.PictureSupport.pH264Support->MaxL0ReferencesForP, |
|
support.PictureSupport.pH264Support->MaxL1ReferencesForB); |
|
ref_l1 = support.PictureSupport.pH264Support->MaxL1ReferencesForB; |
|
break; |
|
|
|
case D3D12_VIDEO_ENCODER_CODEC_HEVC: |
|
ref_l0 = FFMIN(support.PictureSupport.pHEVCSupport->MaxL0ReferencesForP, |
|
support.PictureSupport.pHEVCSupport->MaxL1ReferencesForB); |
|
ref_l1 = support.PictureSupport.pHEVCSupport->MaxL1ReferencesForB; |
|
break; |
|
|
|
default: |
|
av_assert0(0); |
|
} |
|
} else { |
|
ref_l0 = ref_l1 = 0; |
|
} |
|
|
|
if (ref_l0 > 0 && ref_l1 > 0 && ctx->bi_not_empty) { |
|
base_ctx->p_to_gpb = 1; |
|
av_log(avctx, AV_LOG_VERBOSE, "Driver does not support P-frames, " |
|
"replacing them with B-frames.\n"); |
|
} |
|
|
|
err = ff_hw_base_init_gop_structure(base_ctx, avctx, ref_l0, ref_l1, ctx->codec->flags, 0); |
|
if (err < 0) |
|
return err; |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_create_encoder(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
AVD3D12VAFramesContext *frames_hwctx = base_ctx->input_frames->hwctx; |
|
HRESULT hr; |
|
|
|
D3D12_VIDEO_ENCODER_DESC desc = { |
|
.NodeMask = 0, |
|
.Flags = D3D12_VIDEO_ENCODER_FLAG_NONE, |
|
.EncodeCodec = ctx->codec->d3d12_codec, |
|
.EncodeProfile = ctx->profile->d3d12_profile, |
|
.InputFormat = frames_hwctx->format, |
|
.CodecConfiguration = ctx->codec_conf, |
|
.MaxMotionEstimationPrecision = D3D12_VIDEO_ENCODER_MOTION_ESTIMATION_PRECISION_MODE_MAXIMUM, |
|
}; |
|
|
|
hr = ID3D12VideoDevice3_CreateVideoEncoder(ctx->video_device3, &desc, &IID_ID3D12VideoEncoder, |
|
(void **)&ctx->encoder); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create encoder.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_create_encoder_heap(AVCodecContext *avctx) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
HRESULT hr; |
|
|
|
D3D12_VIDEO_ENCODER_HEAP_DESC desc = { |
|
.NodeMask = 0, |
|
.Flags = D3D12_VIDEO_ENCODER_FLAG_NONE, |
|
.EncodeCodec = ctx->codec->d3d12_codec, |
|
.EncodeProfile = ctx->profile->d3d12_profile, |
|
.EncodeLevel = ctx->level, |
|
.ResolutionsListCount = 1, |
|
.pResolutionList = &ctx->resolution, |
|
}; |
|
|
|
hr = ID3D12VideoDevice3_CreateVideoEncoderHeap(ctx->video_device3, &desc, |
|
&IID_ID3D12VideoEncoderHeap, (void **)&ctx->encoder_heap); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create encoder heap.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static void d3d12va_encode_free_buffer(void *opaque, uint8_t *data) |
|
{ |
|
ID3D12Resource *pResource; |
|
|
|
pResource = (ID3D12Resource *)data; |
|
D3D12_OBJECT_RELEASE(pResource); |
|
} |
|
|
|
static AVBufferRef *d3d12va_encode_alloc_output_buffer(void *opaque, size_t size) |
|
{ |
|
AVCodecContext *avctx = opaque; |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
ID3D12Resource *pResource = NULL; |
|
HRESULT hr; |
|
AVBufferRef *ref; |
|
D3D12_HEAP_PROPERTIES heap_props; |
|
D3D12_HEAP_TYPE heap_type = D3D12_HEAP_TYPE_READBACK; |
|
|
|
D3D12_RESOURCE_DESC desc = { |
|
.Dimension = D3D12_RESOURCE_DIMENSION_BUFFER, |
|
.Alignment = 0, |
|
.Width = FFALIGN(3 * base_ctx->surface_width * base_ctx->surface_height + (1 << 16), |
|
D3D12_TEXTURE_DATA_PLACEMENT_ALIGNMENT), |
|
.Height = 1, |
|
.DepthOrArraySize = 1, |
|
.MipLevels = 1, |
|
.Format = DXGI_FORMAT_UNKNOWN, |
|
.SampleDesc = { .Count = 1, .Quality = 0 }, |
|
.Layout = D3D12_TEXTURE_LAYOUT_ROW_MAJOR, |
|
.Flags = D3D12_RESOURCE_FLAG_NONE, |
|
}; |
|
|
|
ctx->hwctx->device->lpVtbl->GetCustomHeapProperties(ctx->hwctx->device, &heap_props, 0, heap_type); |
|
|
|
hr = ID3D12Device_CreateCommittedResource(ctx->hwctx->device, &heap_props, D3D12_HEAP_FLAG_NONE, |
|
&desc, D3D12_RESOURCE_STATE_COMMON, NULL, &IID_ID3D12Resource, |
|
(void **)&pResource); |
|
|
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create d3d12 buffer.\n"); |
|
return NULL; |
|
} |
|
|
|
ref = av_buffer_create((uint8_t *)(uintptr_t)pResource, |
|
sizeof(pResource), |
|
&d3d12va_encode_free_buffer, |
|
avctx, AV_BUFFER_FLAG_READONLY); |
|
if (!ref) { |
|
D3D12_OBJECT_RELEASE(pResource); |
|
return NULL; |
|
} |
|
|
|
return ref; |
|
} |
|
|
|
static int d3d12va_encode_prepare_output_buffers(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
AVD3D12VAFramesContext *frames_ctx = base_ctx->input_frames->hwctx; |
|
HRESULT hr; |
|
|
|
ctx->req.NodeIndex = 0; |
|
ctx->req.Codec = ctx->codec->d3d12_codec; |
|
ctx->req.Profile = ctx->profile->d3d12_profile; |
|
ctx->req.InputFormat = frames_ctx->format; |
|
ctx->req.PictureTargetResolution = ctx->resolution; |
|
|
|
hr = ID3D12VideoDevice3_CheckFeatureSupport(ctx->video_device3, |
|
D3D12_FEATURE_VIDEO_ENCODER_RESOURCE_REQUIREMENTS, |
|
&ctx->req, sizeof(ctx->req)); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to check encoder resource requirements support.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
if (!ctx->req.IsSupported) { |
|
av_log(avctx, AV_LOG_ERROR, "Encoder resource requirements unsupported.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
ctx->output_buffer_pool = av_buffer_pool_init2(sizeof(ID3D12Resource *), avctx, |
|
&d3d12va_encode_alloc_output_buffer, NULL); |
|
if (!ctx->output_buffer_pool) |
|
return AVERROR(ENOMEM); |
|
|
|
return 0; |
|
} |
|
|
|
static int d3d12va_encode_create_command_objects(AVCodecContext *avctx) |
|
{ |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
ID3D12CommandAllocator *command_allocator = NULL; |
|
int err; |
|
HRESULT hr; |
|
|
|
D3D12_COMMAND_QUEUE_DESC queue_desc = { |
|
.Type = D3D12_COMMAND_LIST_TYPE_VIDEO_ENCODE, |
|
.Priority = 0, |
|
.Flags = D3D12_COMMAND_QUEUE_FLAG_NONE, |
|
.NodeMask = 0, |
|
}; |
|
|
|
ctx->allocator_queue = av_fifo_alloc2(D3D12VA_VIDEO_ENC_ASYNC_DEPTH, |
|
sizeof(CommandAllocator), AV_FIFO_FLAG_AUTO_GROW); |
|
if (!ctx->allocator_queue) |
|
return AVERROR(ENOMEM); |
|
|
|
hr = ID3D12Device_CreateFence(ctx->hwctx->device, 0, D3D12_FENCE_FLAG_NONE, |
|
&IID_ID3D12Fence, (void **)&ctx->sync_ctx.fence); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create fence(%lx)\n", (long)hr); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
ctx->sync_ctx.event = CreateEvent(NULL, FALSE, FALSE, NULL); |
|
if (!ctx->sync_ctx.event) |
|
goto fail; |
|
|
|
err = d3d12va_get_valid_command_allocator(avctx, &command_allocator); |
|
if (err < 0) |
|
goto fail; |
|
|
|
hr = ID3D12Device_CreateCommandQueue(ctx->hwctx->device, &queue_desc, |
|
&IID_ID3D12CommandQueue, (void **)&ctx->command_queue); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create command queue(%lx)\n", (long)hr); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12Device_CreateCommandList(ctx->hwctx->device, 0, queue_desc.Type, |
|
command_allocator, NULL, &IID_ID3D12CommandList, |
|
(void **)&ctx->command_list); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to create command list(%lx)\n", (long)hr); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12VideoEncodeCommandList2_Close(ctx->command_list); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to close the command list(%lx)\n", (long)hr); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
ID3D12CommandQueue_ExecuteCommandLists(ctx->command_queue, 1, (ID3D12CommandList **)&ctx->command_list); |
|
|
|
err = d3d12va_sync_with_gpu(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
err = d3d12va_discard_command_allocator(avctx, command_allocator, ctx->sync_ctx.fence_value); |
|
if (err < 0) |
|
goto fail; |
|
|
|
return 0; |
|
|
|
fail: |
|
D3D12_OBJECT_RELEASE(command_allocator); |
|
return err; |
|
} |
|
|
|
static int d3d12va_encode_create_recon_frames(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
AVD3D12VAFramesContext *hwctx; |
|
enum AVPixelFormat recon_format; |
|
int err; |
|
|
|
err = ff_hw_base_get_recon_format(base_ctx, NULL, &recon_format); |
|
if (err < 0) |
|
return err; |
|
|
|
base_ctx->recon_frames_ref = av_hwframe_ctx_alloc(base_ctx->device_ref); |
|
if (!base_ctx->recon_frames_ref) |
|
return AVERROR(ENOMEM); |
|
|
|
base_ctx->recon_frames = (AVHWFramesContext *)base_ctx->recon_frames_ref->data; |
|
hwctx = (AVD3D12VAFramesContext *)base_ctx->recon_frames->hwctx; |
|
|
|
base_ctx->recon_frames->format = AV_PIX_FMT_D3D12; |
|
base_ctx->recon_frames->sw_format = recon_format; |
|
base_ctx->recon_frames->width = base_ctx->surface_width; |
|
base_ctx->recon_frames->height = base_ctx->surface_height; |
|
|
|
hwctx->flags = D3D12_RESOURCE_FLAG_VIDEO_ENCODE_REFERENCE_ONLY | |
|
D3D12_RESOURCE_FLAG_DENY_SHADER_RESOURCE; |
|
|
|
err = av_hwframe_ctx_init(base_ctx->recon_frames_ref); |
|
if (err < 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Failed to initialise reconstructed " |
|
"frame context: %d.\n", err); |
|
return err; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static const FFHWEncodePictureOperation d3d12va_type = { |
|
.priv_size = sizeof(D3D12VAEncodePicture), |
|
|
|
.init = &d3d12va_encode_init, |
|
|
|
.issue = &d3d12va_encode_issue, |
|
|
|
.output = &d3d12va_encode_output, |
|
|
|
.free = &d3d12va_encode_free, |
|
}; |
|
|
|
int ff_d3d12va_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt) |
|
{ |
|
return ff_hw_base_encode_receive_packet(avctx->priv_data, avctx, pkt); |
|
} |
|
|
|
int ff_d3d12va_encode_init(AVCodecContext *avctx) |
|
{ |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
D3D12_FEATURE_DATA_VIDEO_FEATURE_AREA_SUPPORT support = { 0 }; |
|
int err; |
|
HRESULT hr; |
|
|
|
err = ff_hw_base_encode_init(avctx, base_ctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
base_ctx->op = &d3d12va_type; |
|
|
|
ctx->hwctx = base_ctx->device->hwctx; |
|
|
|
ctx->resolution.Width = base_ctx->input_frames->width; |
|
ctx->resolution.Height = base_ctx->input_frames->height; |
|
|
|
hr = ID3D12Device_QueryInterface(ctx->hwctx->device, &IID_ID3D12Device3, (void **)&ctx->device3); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "ID3D12Device3 interface is not supported.\n"); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
hr = ID3D12Device3_QueryInterface(ctx->device3, &IID_ID3D12VideoDevice3, (void **)&ctx->video_device3); |
|
if (FAILED(hr)) { |
|
av_log(avctx, AV_LOG_ERROR, "ID3D12VideoDevice3 interface is not supported.\n"); |
|
err = AVERROR_UNKNOWN; |
|
goto fail; |
|
} |
|
|
|
if (FAILED(ID3D12VideoDevice3_CheckFeatureSupport(ctx->video_device3, D3D12_FEATURE_VIDEO_FEATURE_AREA_SUPPORT, |
|
&support, sizeof(support))) && !support.VideoEncodeSupport) { |
|
av_log(avctx, AV_LOG_ERROR, "D3D12 video device has no video encoder support.\n"); |
|
err = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
err = d3d12va_encode_set_profile(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
err = d3d12va_encode_init_rate_control(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
if (ctx->codec->get_encoder_caps) { |
|
err = ctx->codec->get_encoder_caps(avctx); |
|
if (err < 0) |
|
goto fail; |
|
} |
|
|
|
err = d3d12va_encode_init_gop_structure(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
if (!(ctx->codec->flags & FF_HW_FLAG_SLICE_CONTROL) && avctx->slices > 0) { |
|
av_log(avctx, AV_LOG_WARNING, "Multiple slices were requested " |
|
"but this codec does not support controlling slices.\n"); |
|
} |
|
|
|
err = d3d12va_encode_create_command_objects(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
err = d3d12va_encode_create_recon_frames(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
err = d3d12va_encode_prepare_output_buffers(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
if (ctx->codec->configure) { |
|
err = ctx->codec->configure(avctx); |
|
if (err < 0) |
|
goto fail; |
|
} |
|
|
|
if (ctx->codec->init_sequence_params) { |
|
err = ctx->codec->init_sequence_params(avctx); |
|
if (err < 0) { |
|
av_log(avctx, AV_LOG_ERROR, "Codec sequence initialisation " |
|
"failed: %d.\n", err); |
|
goto fail; |
|
} |
|
} |
|
|
|
if (ctx->codec->set_level) { |
|
err = ctx->codec->set_level(avctx); |
|
if (err < 0) |
|
goto fail; |
|
} |
|
|
|
base_ctx->output_delay = base_ctx->b_per_p; |
|
base_ctx->decode_delay = base_ctx->max_b_depth; |
|
|
|
err = d3d12va_create_encoder(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
err = d3d12va_create_encoder_heap(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
base_ctx->async_encode = 1; |
|
base_ctx->encode_fifo = av_fifo_alloc2(base_ctx->async_depth, |
|
sizeof(D3D12VAEncodePicture *), 0); |
|
if (!base_ctx->encode_fifo) |
|
return AVERROR(ENOMEM); |
|
|
|
return 0; |
|
|
|
fail: |
|
return err; |
|
} |
|
|
|
int ff_d3d12va_encode_close(AVCodecContext *avctx) |
|
{ |
|
int num_allocator = 0; |
|
FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
|
D3D12VAEncodeContext *ctx = avctx->priv_data; |
|
FFHWBaseEncodePicture *pic, *next; |
|
CommandAllocator allocator; |
|
|
|
if (!base_ctx->frame) |
|
return 0; |
|
|
|
for (pic = base_ctx->pic_start; pic; pic = next) { |
|
next = pic->next; |
|
d3d12va_encode_free(avctx, pic); |
|
} |
|
|
|
d3d12va_encode_free_rc_params(avctx); |
|
|
|
av_buffer_pool_uninit(&ctx->output_buffer_pool); |
|
|
|
D3D12_OBJECT_RELEASE(ctx->command_list); |
|
D3D12_OBJECT_RELEASE(ctx->command_queue); |
|
|
|
if (ctx->allocator_queue) { |
|
while (av_fifo_read(ctx->allocator_queue, &allocator, 1) >= 0) { |
|
num_allocator++; |
|
D3D12_OBJECT_RELEASE(allocator.command_allocator); |
|
} |
|
|
|
av_log(avctx, AV_LOG_VERBOSE, "Total number of command allocators reused: %d\n", num_allocator); |
|
} |
|
|
|
av_fifo_freep2(&ctx->allocator_queue); |
|
|
|
D3D12_OBJECT_RELEASE(ctx->sync_ctx.fence); |
|
if (ctx->sync_ctx.event) |
|
CloseHandle(ctx->sync_ctx.event); |
|
|
|
D3D12_OBJECT_RELEASE(ctx->encoder_heap); |
|
D3D12_OBJECT_RELEASE(ctx->encoder); |
|
D3D12_OBJECT_RELEASE(ctx->video_device3); |
|
D3D12_OBJECT_RELEASE(ctx->device3); |
|
|
|
ff_hw_base_encode_close(base_ctx); |
|
|
|
return 0; |
|
}
|
|
|