mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1404 lines
47 KiB
1404 lines
47 KiB
/* |
|
* Videotoolbox hardware acceleration |
|
* |
|
* copyright (c) 2012 Sebastien Zwickert |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include "config.h" |
|
#include "videotoolbox.h" |
|
#include "libavutil/hwcontext_videotoolbox.h" |
|
#include "vt_internal.h" |
|
#include "libavutil/avutil.h" |
|
#include "libavutil/hwcontext.h" |
|
#include "libavutil/pixdesc.h" |
|
#include "bytestream.h" |
|
#include "decode.h" |
|
#include "h264dec.h" |
|
#include "hevcdec.h" |
|
#include "mpegvideo.h" |
|
#include "proresdec.h" |
|
#include <Availability.h> |
|
#include <AvailabilityMacros.h> |
|
#include <TargetConditionals.h> |
|
|
|
#ifndef kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder |
|
# define kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder CFSTR("RequireHardwareAcceleratedVideoDecoder") |
|
#endif |
|
#ifndef kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder |
|
# define kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder CFSTR("EnableHardwareAcceleratedVideoDecoder") |
|
#endif |
|
|
|
#if !HAVE_KCMVIDEOCODECTYPE_HEVC |
|
enum { kCMVideoCodecType_HEVC = 'hvc1' }; |
|
#endif |
|
|
|
#if !HAVE_KCMVIDEOCODECTYPE_VP9 |
|
enum { kCMVideoCodecType_VP9 = 'vp09' }; |
|
#endif |
|
|
|
#define VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING 12 |
|
|
|
typedef struct VTHWFrame { |
|
CVPixelBufferRef pixbuf; |
|
AVBufferRef *hw_frames_ctx; |
|
} VTHWFrame; |
|
|
|
static void videotoolbox_buffer_release(void *opaque, uint8_t *data) |
|
{ |
|
VTHWFrame *ref = (VTHWFrame *)data; |
|
av_buffer_unref(&ref->hw_frames_ctx); |
|
CVPixelBufferRelease(ref->pixbuf); |
|
|
|
av_free(data); |
|
} |
|
|
|
int ff_videotoolbox_buffer_copy(VTContext *vtctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
void *tmp; |
|
|
|
tmp = av_fast_realloc(vtctx->bitstream, |
|
&vtctx->allocated_size, |
|
size); |
|
|
|
if (!tmp) |
|
return AVERROR(ENOMEM); |
|
|
|
vtctx->bitstream = tmp; |
|
memcpy(vtctx->bitstream, buffer, size); |
|
vtctx->bitstream_size = size; |
|
|
|
return 0; |
|
} |
|
|
|
static int videotoolbox_postproc_frame(void *avctx, AVFrame *frame) |
|
{ |
|
int ret; |
|
VTHWFrame *ref = (VTHWFrame *)frame->buf[0]->data; |
|
|
|
if (!ref->pixbuf) { |
|
av_log(avctx, AV_LOG_ERROR, "No frame decoded?\n"); |
|
av_frame_unref(frame); |
|
return AVERROR_EXTERNAL; |
|
} |
|
|
|
frame->crop_right = 0; |
|
frame->crop_left = 0; |
|
frame->crop_top = 0; |
|
frame->crop_bottom = 0; |
|
|
|
if ((ret = av_vt_pixbuf_set_attachments(avctx, ref->pixbuf, frame)) < 0) |
|
return ret; |
|
|
|
frame->data[3] = (uint8_t*)ref->pixbuf; |
|
|
|
if (ref->hw_frames_ctx) { |
|
av_buffer_unref(&frame->hw_frames_ctx); |
|
frame->hw_frames_ctx = av_buffer_ref(ref->hw_frames_ctx); |
|
if (!frame->hw_frames_ctx) |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
int ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame) |
|
{ |
|
size_t size = sizeof(VTHWFrame); |
|
uint8_t *data = NULL; |
|
AVBufferRef *buf = NULL; |
|
int ret = ff_attach_decode_data(frame); |
|
FrameDecodeData *fdd; |
|
if (ret < 0) |
|
return ret; |
|
|
|
data = av_mallocz(size); |
|
if (!data) |
|
return AVERROR(ENOMEM); |
|
buf = av_buffer_create(data, size, videotoolbox_buffer_release, NULL, 0); |
|
if (!buf) { |
|
av_freep(&data); |
|
return AVERROR(ENOMEM); |
|
} |
|
frame->buf[0] = buf; |
|
|
|
fdd = (FrameDecodeData*)frame->private_ref->data; |
|
fdd->post_process = videotoolbox_postproc_frame; |
|
|
|
frame->width = avctx->width; |
|
frame->height = avctx->height; |
|
frame->format = avctx->pix_fmt; |
|
|
|
return 0; |
|
} |
|
|
|
#define AV_W8(p, v) *(p) = (v) |
|
|
|
static int escape_ps(uint8_t* dst, const uint8_t* src, int src_size) |
|
{ |
|
int i; |
|
int size = src_size; |
|
uint8_t* p = dst; |
|
|
|
for (i = 0; i < src_size; i++) { |
|
if (i + 2 < src_size && |
|
src[i] == 0x00 && |
|
src[i + 1] == 0x00 && |
|
src[i + 2] <= 0x03) { |
|
if (dst) { |
|
*p++ = src[i++]; |
|
*p++ = src[i++]; |
|
*p++ = 0x03; |
|
} else { |
|
i += 2; |
|
} |
|
size++; |
|
} |
|
if (dst) |
|
*p++ = src[i]; |
|
} |
|
|
|
if (dst) |
|
av_assert0((p - dst) == size); |
|
|
|
return size; |
|
} |
|
|
|
CFDataRef ff_videotoolbox_avcc_extradata_create(AVCodecContext *avctx) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
H264Context *h = avctx->priv_data; |
|
CFDataRef data = NULL; |
|
uint8_t *p; |
|
int sps_size = escape_ps(NULL, h->ps.sps->data, h->ps.sps->data_size); |
|
int pps_size = escape_ps(NULL, h->ps.pps->data, h->ps.pps->data_size); |
|
int vt_extradata_size; |
|
uint8_t *vt_extradata; |
|
|
|
vt_extradata_size = 6 + 2 + sps_size + 3 + pps_size; |
|
vt_extradata = av_malloc(vt_extradata_size); |
|
|
|
if (!vt_extradata) |
|
return NULL; |
|
|
|
p = vt_extradata; |
|
|
|
AV_W8(p + 0, 1); /* version */ |
|
AV_W8(p + 1, h->ps.sps->data[1]); /* profile */ |
|
AV_W8(p + 2, h->ps.sps->data[2]); /* profile compat */ |
|
AV_W8(p + 3, h->ps.sps->data[3]); /* level */ |
|
AV_W8(p + 4, 0xff); /* 6 bits reserved (111111) + 2 bits nal size length - 3 (11) */ |
|
AV_W8(p + 5, 0xe1); /* 3 bits reserved (111) + 5 bits number of sps (00001) */ |
|
AV_WB16(p + 6, sps_size); |
|
p += 8; |
|
p += escape_ps(p, h->ps.sps->data, h->ps.sps->data_size); |
|
AV_W8(p + 0, 1); /* number of pps */ |
|
AV_WB16(p + 1, pps_size); |
|
p += 3; |
|
p += escape_ps(p, h->ps.pps->data, h->ps.pps->data_size); |
|
|
|
av_assert0(p - vt_extradata == vt_extradata_size); |
|
|
|
// save sps header (profile/level) used to create decoder session, |
|
// so we can detect changes and recreate it. |
|
if (vtctx) |
|
memcpy(vtctx->sps, h->ps.sps->data + 1, 3); |
|
|
|
data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size); |
|
av_free(vt_extradata); |
|
return data; |
|
} |
|
|
|
CFDataRef ff_videotoolbox_hvcc_extradata_create(AVCodecContext *avctx) |
|
{ |
|
HEVCContext *h = avctx->priv_data; |
|
int i, num_vps = 0, num_sps = 0, num_pps = 0; |
|
const HEVCVPS *vps = h->ps.vps; |
|
const HEVCSPS *sps = h->ps.sps; |
|
const HEVCPPS *pps = h->ps.pps; |
|
PTLCommon ptlc = vps->ptl.general_ptl; |
|
VUI vui = sps->vui; |
|
uint8_t parallelismType; |
|
CFDataRef data = NULL; |
|
uint8_t *p; |
|
int vt_extradata_size = 23 + 3 + 3 + 3; |
|
uint8_t *vt_extradata; |
|
|
|
#define COUNT_SIZE_PS(T, t) \ |
|
for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \ |
|
if (h->ps.t##ps_list[i]) { \ |
|
const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \ |
|
vt_extradata_size += 2 + lps->data_size; \ |
|
num_##t##ps++; \ |
|
} \ |
|
} |
|
|
|
COUNT_SIZE_PS(V, v) |
|
COUNT_SIZE_PS(S, s) |
|
COUNT_SIZE_PS(P, p) |
|
|
|
vt_extradata = av_malloc(vt_extradata_size); |
|
if (!vt_extradata) |
|
return NULL; |
|
p = vt_extradata; |
|
|
|
/* unsigned int(8) configurationVersion = 1; */ |
|
AV_W8(p + 0, 1); |
|
|
|
/* |
|
* unsigned int(2) general_profile_space; |
|
* unsigned int(1) general_tier_flag; |
|
* unsigned int(5) general_profile_idc; |
|
*/ |
|
AV_W8(p + 1, ptlc.profile_space << 6 | |
|
ptlc.tier_flag << 5 | |
|
ptlc.profile_idc); |
|
|
|
/* unsigned int(32) general_profile_compatibility_flags; */ |
|
memcpy(p + 2, ptlc.profile_compatibility_flag, 4); |
|
|
|
/* unsigned int(48) general_constraint_indicator_flags; */ |
|
AV_W8(p + 6, ptlc.progressive_source_flag << 7 | |
|
ptlc.interlaced_source_flag << 6 | |
|
ptlc.non_packed_constraint_flag << 5 | |
|
ptlc.frame_only_constraint_flag << 4); |
|
AV_W8(p + 7, 0); |
|
AV_WN32(p + 8, 0); |
|
|
|
/* unsigned int(8) general_level_idc; */ |
|
AV_W8(p + 12, ptlc.level_idc); |
|
|
|
/* |
|
* bit(4) reserved = ‘1111’b; |
|
* unsigned int(12) min_spatial_segmentation_idc; |
|
*/ |
|
AV_W8(p + 13, 0xf0 | (vui.min_spatial_segmentation_idc >> 4)); |
|
AV_W8(p + 14, vui.min_spatial_segmentation_idc & 0xff); |
|
|
|
/* |
|
* bit(6) reserved = ‘111111’b; |
|
* unsigned int(2) parallelismType; |
|
*/ |
|
if (!vui.min_spatial_segmentation_idc) |
|
parallelismType = 0; |
|
else if (pps->entropy_coding_sync_enabled_flag && pps->tiles_enabled_flag) |
|
parallelismType = 0; |
|
else if (pps->entropy_coding_sync_enabled_flag) |
|
parallelismType = 3; |
|
else if (pps->tiles_enabled_flag) |
|
parallelismType = 2; |
|
else |
|
parallelismType = 1; |
|
AV_W8(p + 15, 0xfc | parallelismType); |
|
|
|
/* |
|
* bit(6) reserved = ‘111111’b; |
|
* unsigned int(2) chromaFormat; |
|
*/ |
|
AV_W8(p + 16, sps->chroma_format_idc | 0xfc); |
|
|
|
/* |
|
* bit(5) reserved = ‘11111’b; |
|
* unsigned int(3) bitDepthLumaMinus8; |
|
*/ |
|
AV_W8(p + 17, (sps->bit_depth - 8) | 0xfc); |
|
|
|
/* |
|
* bit(5) reserved = ‘11111’b; |
|
* unsigned int(3) bitDepthChromaMinus8; |
|
*/ |
|
AV_W8(p + 18, (sps->bit_depth_chroma - 8) | 0xfc); |
|
|
|
/* bit(16) avgFrameRate; */ |
|
AV_WB16(p + 19, 0); |
|
|
|
/* |
|
* bit(2) constantFrameRate; |
|
* bit(3) numTemporalLayers; |
|
* bit(1) temporalIdNested; |
|
* unsigned int(2) lengthSizeMinusOne; |
|
*/ |
|
AV_W8(p + 21, 0 << 6 | |
|
sps->max_sub_layers << 3 | |
|
sps->temporal_id_nesting_flag << 2 | |
|
3); |
|
|
|
/* unsigned int(8) numOfArrays; */ |
|
AV_W8(p + 22, 3); |
|
|
|
p += 23; |
|
|
|
#define APPEND_PS(T, t) \ |
|
/* \ |
|
* bit(1) array_completeness; \ |
|
* unsigned int(1) reserved = 0; \ |
|
* unsigned int(6) NAL_unit_type; \ |
|
*/ \ |
|
AV_W8(p, 1 << 7 | \ |
|
HEVC_NAL_##T##PS & 0x3f); \ |
|
/* unsigned int(16) numNalus; */ \ |
|
AV_WB16(p + 1, num_##t##ps); \ |
|
p += 3; \ |
|
for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \ |
|
if (h->ps.t##ps_list[i]) { \ |
|
const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \ |
|
/* unsigned int(16) nalUnitLength; */ \ |
|
AV_WB16(p, lps->data_size); \ |
|
/* bit(8*nalUnitLength) nalUnit; */ \ |
|
memcpy(p + 2, lps->data, lps->data_size); \ |
|
p += 2 + lps->data_size; \ |
|
} \ |
|
} |
|
|
|
APPEND_PS(V, v) |
|
APPEND_PS(S, s) |
|
APPEND_PS(P, p) |
|
|
|
av_assert0(p - vt_extradata == vt_extradata_size); |
|
|
|
data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size); |
|
av_free(vt_extradata); |
|
return data; |
|
} |
|
|
|
int ff_videotoolbox_h264_start_frame(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
H264Context *h = avctx->priv_data; |
|
|
|
if (h->is_avc == 1) { |
|
return ff_videotoolbox_buffer_copy(vtctx, buffer, size); |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static int videotoolbox_h264_decode_params(AVCodecContext *avctx, |
|
int type, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
H264Context *h = avctx->priv_data; |
|
|
|
// save sps header (profile/level) used to create decoder session |
|
if (!vtctx->sps[0]) |
|
memcpy(vtctx->sps, h->ps.sps->data + 1, 3); |
|
|
|
if (type == H264_NAL_SPS) { |
|
if (size > 4 && memcmp(vtctx->sps, buffer + 1, 3) != 0) { |
|
vtctx->reconfig_needed = true; |
|
memcpy(vtctx->sps, buffer + 1, 3); |
|
} |
|
} |
|
|
|
// pass-through SPS/PPS changes to the decoder |
|
return ff_videotoolbox_h264_decode_slice(avctx, buffer, size); |
|
} |
|
|
|
static int videotoolbox_common_decode_slice(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
void *tmp; |
|
|
|
tmp = av_fast_realloc(vtctx->bitstream, |
|
&vtctx->allocated_size, |
|
vtctx->bitstream_size+size+4); |
|
if (!tmp) |
|
return AVERROR(ENOMEM); |
|
|
|
vtctx->bitstream = tmp; |
|
|
|
AV_WB32(vtctx->bitstream + vtctx->bitstream_size, size); |
|
memcpy(vtctx->bitstream + vtctx->bitstream_size + 4, buffer, size); |
|
|
|
vtctx->bitstream_size += size + 4; |
|
|
|
return 0; |
|
} |
|
|
|
int ff_videotoolbox_h264_decode_slice(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
H264Context *h = avctx->priv_data; |
|
|
|
if (h->is_avc == 1) |
|
return 0; |
|
|
|
return videotoolbox_common_decode_slice(avctx, buffer, size); |
|
} |
|
|
|
#if CONFIG_VIDEOTOOLBOX |
|
// Return the AVVideotoolboxContext that matters currently. Where it comes from |
|
// depends on the API used. |
|
static AVVideotoolboxContext *videotoolbox_get_context(AVCodecContext *avctx) |
|
{ |
|
// Somewhat tricky because the user can call av_videotoolbox_default_free() |
|
// at any time, even when the codec is closed. |
|
if (avctx->internal && avctx->internal->hwaccel_priv_data) { |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
if (vtctx->vt_ctx) |
|
return vtctx->vt_ctx; |
|
} |
|
return avctx->hwaccel_context; |
|
} |
|
|
|
static void videotoolbox_stop(AVCodecContext *avctx) |
|
{ |
|
AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx); |
|
if (!videotoolbox) |
|
return; |
|
|
|
if (videotoolbox->cm_fmt_desc) { |
|
CFRelease(videotoolbox->cm_fmt_desc); |
|
videotoolbox->cm_fmt_desc = NULL; |
|
} |
|
|
|
if (videotoolbox->session) { |
|
VTDecompressionSessionInvalidate(videotoolbox->session); |
|
CFRelease(videotoolbox->session); |
|
videotoolbox->session = NULL; |
|
} |
|
} |
|
|
|
int ff_videotoolbox_uninit(AVCodecContext *avctx) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
if (!vtctx) |
|
return 0; |
|
|
|
av_freep(&vtctx->bitstream); |
|
if (vtctx->frame) |
|
CVPixelBufferRelease(vtctx->frame); |
|
|
|
if (vtctx->vt_ctx) |
|
videotoolbox_stop(avctx); |
|
|
|
av_buffer_unref(&vtctx->cached_hw_frames_ctx); |
|
av_freep(&vtctx->vt_ctx); |
|
|
|
return 0; |
|
} |
|
|
|
static int videotoolbox_buffer_create(AVCodecContext *avctx, AVFrame *frame) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
CVPixelBufferRef pixbuf = (CVPixelBufferRef)vtctx->frame; |
|
OSType pixel_format = CVPixelBufferGetPixelFormatType(pixbuf); |
|
enum AVPixelFormat sw_format = av_map_videotoolbox_format_to_pixfmt(pixel_format); |
|
int width = CVPixelBufferGetWidth(pixbuf); |
|
int height = CVPixelBufferGetHeight(pixbuf); |
|
AVHWFramesContext *cached_frames; |
|
VTHWFrame *ref; |
|
int ret; |
|
|
|
if (!frame->buf[0] || frame->data[3]) { |
|
av_log(avctx, AV_LOG_ERROR, "videotoolbox: invalid state\n"); |
|
av_frame_unref(frame); |
|
return AVERROR_EXTERNAL; |
|
} |
|
|
|
ref = (VTHWFrame *)frame->buf[0]->data; |
|
|
|
if (ref->pixbuf) |
|
CVPixelBufferRelease(ref->pixbuf); |
|
ref->pixbuf = vtctx->frame; |
|
vtctx->frame = NULL; |
|
|
|
// Old API code path. |
|
if (!vtctx->cached_hw_frames_ctx) |
|
return 0; |
|
|
|
cached_frames = (AVHWFramesContext*)vtctx->cached_hw_frames_ctx->data; |
|
|
|
if (cached_frames->sw_format != sw_format || |
|
cached_frames->width != width || |
|
cached_frames->height != height) { |
|
AVBufferRef *hw_frames_ctx = av_hwframe_ctx_alloc(cached_frames->device_ref); |
|
AVHWFramesContext *hw_frames; |
|
if (!hw_frames_ctx) |
|
return AVERROR(ENOMEM); |
|
|
|
hw_frames = (AVHWFramesContext*)hw_frames_ctx->data; |
|
hw_frames->format = cached_frames->format; |
|
hw_frames->sw_format = sw_format; |
|
hw_frames->width = width; |
|
hw_frames->height = height; |
|
|
|
ret = av_hwframe_ctx_init(hw_frames_ctx); |
|
if (ret < 0) { |
|
av_buffer_unref(&hw_frames_ctx); |
|
return ret; |
|
} |
|
|
|
av_buffer_unref(&vtctx->cached_hw_frames_ctx); |
|
vtctx->cached_hw_frames_ctx = hw_frames_ctx; |
|
} |
|
|
|
av_buffer_unref(&ref->hw_frames_ctx); |
|
ref->hw_frames_ctx = av_buffer_ref(vtctx->cached_hw_frames_ctx); |
|
if (!ref->hw_frames_ctx) |
|
return AVERROR(ENOMEM); |
|
|
|
return 0; |
|
} |
|
|
|
static void videotoolbox_write_mp4_descr_length(PutByteContext *pb, int length) |
|
{ |
|
int i; |
|
uint8_t b; |
|
|
|
for (i = 3; i >= 0; i--) { |
|
b = (length >> (i * 7)) & 0x7F; |
|
if (i != 0) |
|
b |= 0x80; |
|
|
|
bytestream2_put_byteu(pb, b); |
|
} |
|
} |
|
|
|
static CFDataRef videotoolbox_esds_extradata_create(AVCodecContext *avctx) |
|
{ |
|
CFDataRef data; |
|
uint8_t *rw_extradata; |
|
PutByteContext pb; |
|
int full_size = 3 + 5 + 13 + 5 + avctx->extradata_size + 3; |
|
// ES_DescrTag data + DecoderConfigDescrTag + data + DecSpecificInfoTag + size + SLConfigDescriptor |
|
int config_size = 13 + 5 + avctx->extradata_size; |
|
int s; |
|
|
|
if (!(rw_extradata = av_mallocz(full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING))) |
|
return NULL; |
|
|
|
bytestream2_init_writer(&pb, rw_extradata, full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING); |
|
bytestream2_put_byteu(&pb, 0); // version |
|
bytestream2_put_ne24(&pb, 0); // flags |
|
|
|
// elementary stream descriptor |
|
bytestream2_put_byteu(&pb, 0x03); // ES_DescrTag |
|
videotoolbox_write_mp4_descr_length(&pb, full_size); |
|
bytestream2_put_ne16(&pb, 0); // esid |
|
bytestream2_put_byteu(&pb, 0); // stream priority (0-32) |
|
|
|
// decoder configuration descriptor |
|
bytestream2_put_byteu(&pb, 0x04); // DecoderConfigDescrTag |
|
videotoolbox_write_mp4_descr_length(&pb, config_size); |
|
bytestream2_put_byteu(&pb, 32); // object type indication. 32 = AV_CODEC_ID_MPEG4 |
|
bytestream2_put_byteu(&pb, 0x11); // stream type |
|
bytestream2_put_ne24(&pb, 0); // buffer size |
|
bytestream2_put_ne32(&pb, 0); // max bitrate |
|
bytestream2_put_ne32(&pb, 0); // avg bitrate |
|
|
|
// decoder specific descriptor |
|
bytestream2_put_byteu(&pb, 0x05); ///< DecSpecificInfoTag |
|
videotoolbox_write_mp4_descr_length(&pb, avctx->extradata_size); |
|
|
|
bytestream2_put_buffer(&pb, avctx->extradata, avctx->extradata_size); |
|
|
|
// SLConfigDescriptor |
|
bytestream2_put_byteu(&pb, 0x06); // SLConfigDescrTag |
|
bytestream2_put_byteu(&pb, 0x01); // length |
|
bytestream2_put_byteu(&pb, 0x02); // |
|
|
|
s = bytestream2_size_p(&pb); |
|
|
|
data = CFDataCreate(kCFAllocatorDefault, rw_extradata, s); |
|
|
|
av_freep(&rw_extradata); |
|
return data; |
|
} |
|
|
|
static CMSampleBufferRef videotoolbox_sample_buffer_create(CMFormatDescriptionRef fmt_desc, |
|
void *buffer, |
|
int size) |
|
{ |
|
OSStatus status; |
|
CMBlockBufferRef block_buf; |
|
CMSampleBufferRef sample_buf; |
|
|
|
block_buf = NULL; |
|
sample_buf = NULL; |
|
|
|
status = CMBlockBufferCreateWithMemoryBlock(kCFAllocatorDefault,// structureAllocator |
|
buffer, // memoryBlock |
|
size, // blockLength |
|
kCFAllocatorNull, // blockAllocator |
|
NULL, // customBlockSource |
|
0, // offsetToData |
|
size, // dataLength |
|
0, // flags |
|
&block_buf); |
|
|
|
if (!status) { |
|
status = CMSampleBufferCreate(kCFAllocatorDefault, // allocator |
|
block_buf, // dataBuffer |
|
TRUE, // dataReady |
|
0, // makeDataReadyCallback |
|
0, // makeDataReadyRefcon |
|
fmt_desc, // formatDescription |
|
1, // numSamples |
|
0, // numSampleTimingEntries |
|
NULL, // sampleTimingArray |
|
0, // numSampleSizeEntries |
|
NULL, // sampleSizeArray |
|
&sample_buf); |
|
} |
|
|
|
if (block_buf) |
|
CFRelease(block_buf); |
|
|
|
return sample_buf; |
|
} |
|
|
|
static void videotoolbox_decoder_callback(void *opaque, |
|
void *sourceFrameRefCon, |
|
OSStatus status, |
|
VTDecodeInfoFlags flags, |
|
CVImageBufferRef image_buffer, |
|
CMTime pts, |
|
CMTime duration) |
|
{ |
|
AVCodecContext *avctx = opaque; |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
if (vtctx->frame) { |
|
CVPixelBufferRelease(vtctx->frame); |
|
vtctx->frame = NULL; |
|
} |
|
|
|
if (!image_buffer) { |
|
av_log(avctx, AV_LOG_DEBUG, "vt decoder cb: output image buffer is null\n"); |
|
return; |
|
} |
|
|
|
vtctx->frame = CVPixelBufferRetain(image_buffer); |
|
} |
|
|
|
static OSStatus videotoolbox_session_decode_frame(AVCodecContext *avctx) |
|
{ |
|
OSStatus status; |
|
CMSampleBufferRef sample_buf; |
|
AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx); |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
sample_buf = videotoolbox_sample_buffer_create(videotoolbox->cm_fmt_desc, |
|
vtctx->bitstream, |
|
vtctx->bitstream_size); |
|
|
|
if (!sample_buf) |
|
return -1; |
|
|
|
status = VTDecompressionSessionDecodeFrame(videotoolbox->session, |
|
sample_buf, |
|
0, // decodeFlags |
|
NULL, // sourceFrameRefCon |
|
0); // infoFlagsOut |
|
if (status == noErr) |
|
status = VTDecompressionSessionWaitForAsynchronousFrames(videotoolbox->session); |
|
|
|
CFRelease(sample_buf); |
|
|
|
return status; |
|
} |
|
|
|
static CMVideoFormatDescriptionRef videotoolbox_format_desc_create(CMVideoCodecType codec_type, |
|
CFDictionaryRef decoder_spec, |
|
int width, |
|
int height) |
|
{ |
|
CMFormatDescriptionRef cm_fmt_desc; |
|
OSStatus status; |
|
|
|
status = CMVideoFormatDescriptionCreate(kCFAllocatorDefault, |
|
codec_type, |
|
width, |
|
height, |
|
decoder_spec, // Dictionary of extension |
|
&cm_fmt_desc); |
|
|
|
if (status) |
|
return NULL; |
|
|
|
return cm_fmt_desc; |
|
} |
|
|
|
static CFDictionaryRef videotoolbox_buffer_attributes_create(int width, |
|
int height, |
|
OSType pix_fmt) |
|
{ |
|
CFMutableDictionaryRef buffer_attributes; |
|
CFMutableDictionaryRef io_surface_properties; |
|
CFNumberRef cv_pix_fmt; |
|
CFNumberRef w; |
|
CFNumberRef h; |
|
|
|
w = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &width); |
|
h = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &height); |
|
cv_pix_fmt = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &pix_fmt); |
|
|
|
buffer_attributes = CFDictionaryCreateMutable(kCFAllocatorDefault, |
|
4, |
|
&kCFTypeDictionaryKeyCallBacks, |
|
&kCFTypeDictionaryValueCallBacks); |
|
io_surface_properties = CFDictionaryCreateMutable(kCFAllocatorDefault, |
|
0, |
|
&kCFTypeDictionaryKeyCallBacks, |
|
&kCFTypeDictionaryValueCallBacks); |
|
|
|
if (pix_fmt) |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferPixelFormatTypeKey, cv_pix_fmt); |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfacePropertiesKey, io_surface_properties); |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferWidthKey, w); |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferHeightKey, h); |
|
#if TARGET_OS_IPHONE |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferOpenGLESCompatibilityKey, kCFBooleanTrue); |
|
#else |
|
CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfaceOpenGLTextureCompatibilityKey, kCFBooleanTrue); |
|
#endif |
|
|
|
CFRelease(io_surface_properties); |
|
CFRelease(cv_pix_fmt); |
|
CFRelease(w); |
|
CFRelease(h); |
|
|
|
return buffer_attributes; |
|
} |
|
|
|
static CFDictionaryRef videotoolbox_decoder_config_create(CMVideoCodecType codec_type, |
|
AVCodecContext *avctx) |
|
{ |
|
CFMutableDictionaryRef config_info = CFDictionaryCreateMutable(kCFAllocatorDefault, |
|
0, |
|
&kCFTypeDictionaryKeyCallBacks, |
|
&kCFTypeDictionaryValueCallBacks); |
|
|
|
CFDictionarySetValue(config_info, |
|
codec_type == kCMVideoCodecType_HEVC ? |
|
kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder : |
|
kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder, |
|
kCFBooleanTrue); |
|
|
|
CFMutableDictionaryRef avc_info; |
|
CFDataRef data = NULL; |
|
|
|
avc_info = CFDictionaryCreateMutable(kCFAllocatorDefault, |
|
1, |
|
&kCFTypeDictionaryKeyCallBacks, |
|
&kCFTypeDictionaryValueCallBacks); |
|
|
|
switch (codec_type) { |
|
case kCMVideoCodecType_MPEG4Video : |
|
if (avctx->extradata_size) |
|
data = videotoolbox_esds_extradata_create(avctx); |
|
if (data) |
|
CFDictionarySetValue(avc_info, CFSTR("esds"), data); |
|
break; |
|
case kCMVideoCodecType_H264 : |
|
data = ff_videotoolbox_avcc_extradata_create(avctx); |
|
if (data) |
|
CFDictionarySetValue(avc_info, CFSTR("avcC"), data); |
|
break; |
|
case kCMVideoCodecType_HEVC : |
|
data = ff_videotoolbox_hvcc_extradata_create(avctx); |
|
if (data) |
|
CFDictionarySetValue(avc_info, CFSTR("hvcC"), data); |
|
break; |
|
case kCMVideoCodecType_VP9 : |
|
data = ff_videotoolbox_vpcc_extradata_create(avctx); |
|
if (data) |
|
CFDictionarySetValue(avc_info, CFSTR("vpcC"), data); |
|
break; |
|
default: |
|
break; |
|
} |
|
|
|
CFDictionarySetValue(config_info, |
|
kCMFormatDescriptionExtension_SampleDescriptionExtensionAtoms, |
|
avc_info); |
|
|
|
if (data) |
|
CFRelease(data); |
|
|
|
CFRelease(avc_info); |
|
return config_info; |
|
} |
|
|
|
static int videotoolbox_start(AVCodecContext *avctx) |
|
{ |
|
AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx); |
|
OSStatus status; |
|
VTDecompressionOutputCallbackRecord decoder_cb; |
|
CFDictionaryRef decoder_spec; |
|
CFDictionaryRef buf_attr; |
|
|
|
if (!videotoolbox) { |
|
av_log(avctx, AV_LOG_ERROR, "hwaccel context is not set\n"); |
|
return -1; |
|
} |
|
|
|
switch( avctx->codec_id ) { |
|
case AV_CODEC_ID_H263 : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_H263; |
|
break; |
|
case AV_CODEC_ID_H264 : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_H264; |
|
break; |
|
case AV_CODEC_ID_HEVC : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_HEVC; |
|
break; |
|
case AV_CODEC_ID_MPEG1VIDEO : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG1Video; |
|
break; |
|
case AV_CODEC_ID_MPEG2VIDEO : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG2Video; |
|
break; |
|
case AV_CODEC_ID_MPEG4 : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG4Video; |
|
break; |
|
case AV_CODEC_ID_PRORES : |
|
switch (avctx->codec_tag) { |
|
default: |
|
av_log(avctx, AV_LOG_WARNING, "Unknown prores profile %d\n", avctx->codec_tag); |
|
// fall-through |
|
case MKTAG('a','p','c','o'): // kCMVideoCodecType_AppleProRes422Proxy |
|
case MKTAG('a','p','c','s'): // kCMVideoCodecType_AppleProRes422LT |
|
case MKTAG('a','p','c','n'): // kCMVideoCodecType_AppleProRes422 |
|
case MKTAG('a','p','c','h'): // kCMVideoCodecType_AppleProRes422HQ |
|
case MKTAG('a','p','4','h'): // kCMVideoCodecType_AppleProRes4444 |
|
case MKTAG('a','p','4','x'): // kCMVideoCodecType_AppleProRes4444XQ |
|
videotoolbox->cm_codec_type = av_bswap32(avctx->codec_tag); |
|
break; |
|
} |
|
break; |
|
case AV_CODEC_ID_VP9 : |
|
videotoolbox->cm_codec_type = kCMVideoCodecType_VP9; |
|
break; |
|
default : |
|
break; |
|
} |
|
|
|
#if defined(MAC_OS_X_VERSION_10_9) && !TARGET_OS_IPHONE && (MAC_OS_X_VERSION_MAX_ALLOWED >= MAC_OS_X_VERSION_10_9) && AV_HAS_BUILTIN(__builtin_available) |
|
if (avctx->codec_id == AV_CODEC_ID_PRORES) { |
|
if (__builtin_available(macOS 10.9, *)) { |
|
VTRegisterProfessionalVideoWorkflowVideoDecoders(); |
|
} |
|
} |
|
#endif |
|
|
|
#if defined(MAC_OS_VERSION_11_0) && !TARGET_OS_IPHONE && (MAC_OS_X_VERSION_MAX_ALLOWED >= MAC_OS_VERSION_11_0) && AV_HAS_BUILTIN(__builtin_available) |
|
if (__builtin_available(macOS 11.0, *)) { |
|
VTRegisterSupplementalVideoDecoderIfAvailable(videotoolbox->cm_codec_type); |
|
} |
|
#endif |
|
|
|
decoder_spec = videotoolbox_decoder_config_create(videotoolbox->cm_codec_type, avctx); |
|
|
|
if (!decoder_spec) { |
|
av_log(avctx, AV_LOG_ERROR, "decoder specification creation failed\n"); |
|
return -1; |
|
} |
|
|
|
videotoolbox->cm_fmt_desc = videotoolbox_format_desc_create(videotoolbox->cm_codec_type, |
|
decoder_spec, |
|
avctx->width, |
|
avctx->height); |
|
if (!videotoolbox->cm_fmt_desc) { |
|
if (decoder_spec) |
|
CFRelease(decoder_spec); |
|
|
|
av_log(avctx, AV_LOG_ERROR, "format description creation failed\n"); |
|
return -1; |
|
} |
|
|
|
buf_attr = videotoolbox_buffer_attributes_create(avctx->width, |
|
avctx->height, |
|
videotoolbox->cv_pix_fmt_type); |
|
|
|
decoder_cb.decompressionOutputCallback = videotoolbox_decoder_callback; |
|
decoder_cb.decompressionOutputRefCon = avctx; |
|
|
|
status = VTDecompressionSessionCreate(NULL, // allocator |
|
videotoolbox->cm_fmt_desc, // videoFormatDescription |
|
decoder_spec, // videoDecoderSpecification |
|
buf_attr, // destinationImageBufferAttributes |
|
&decoder_cb, // outputCallback |
|
&videotoolbox->session); // decompressionSessionOut |
|
|
|
if (decoder_spec) |
|
CFRelease(decoder_spec); |
|
if (buf_attr) |
|
CFRelease(buf_attr); |
|
|
|
switch (status) { |
|
case kVTVideoDecoderNotAvailableNowErr: |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox session not available.\n"); |
|
return AVERROR(ENOSYS); |
|
case kVTVideoDecoderUnsupportedDataFormatErr: |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox does not support this format.\n"); |
|
return AVERROR(ENOSYS); |
|
case kVTCouldNotFindVideoDecoderErr: |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder for this format not found.\n"); |
|
return AVERROR(ENOSYS); |
|
case kVTVideoDecoderMalfunctionErr: |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox malfunction.\n"); |
|
return AVERROR(EINVAL); |
|
case kVTVideoDecoderBadDataErr: |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox reported invalid data.\n"); |
|
return AVERROR_INVALIDDATA; |
|
case 0: |
|
return 0; |
|
default: |
|
av_log(avctx, AV_LOG_VERBOSE, "Unknown VideoToolbox session creation error %d\n", (int)status); |
|
return AVERROR_UNKNOWN; |
|
} |
|
} |
|
|
|
static const char *videotoolbox_error_string(OSStatus status) |
|
{ |
|
switch (status) { |
|
case kVTVideoDecoderBadDataErr: |
|
return "bad data"; |
|
case kVTVideoDecoderMalfunctionErr: |
|
return "decoder malfunction"; |
|
case kVTInvalidSessionErr: |
|
return "invalid session"; |
|
} |
|
return "unknown"; |
|
} |
|
|
|
int ff_videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame) |
|
{ |
|
OSStatus status; |
|
AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx); |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
if (vtctx->reconfig_needed == true) { |
|
vtctx->reconfig_needed = false; |
|
av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder needs reconfig, restarting..\n"); |
|
videotoolbox_stop(avctx); |
|
if (videotoolbox_start(avctx) != 0) { |
|
return AVERROR_EXTERNAL; |
|
} |
|
} |
|
|
|
if (!videotoolbox->session || !vtctx->bitstream || !vtctx->bitstream_size) |
|
return AVERROR_INVALIDDATA; |
|
|
|
status = videotoolbox_session_decode_frame(avctx); |
|
if (status != noErr) { |
|
if (status == kVTVideoDecoderMalfunctionErr || status == kVTInvalidSessionErr) |
|
vtctx->reconfig_needed = true; |
|
av_log(avctx, AV_LOG_ERROR, "Failed to decode frame (%s, %d)\n", videotoolbox_error_string(status), (int)status); |
|
return AVERROR_UNKNOWN; |
|
} |
|
|
|
if (!vtctx->frame) { |
|
vtctx->reconfig_needed = true; |
|
return AVERROR_UNKNOWN; |
|
} |
|
|
|
return videotoolbox_buffer_create(avctx, frame); |
|
} |
|
|
|
static int videotoolbox_h264_end_frame(AVCodecContext *avctx) |
|
{ |
|
H264Context *h = avctx->priv_data; |
|
AVFrame *frame = h->cur_pic_ptr->f; |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
int ret = ff_videotoolbox_common_end_frame(avctx, frame); |
|
vtctx->bitstream_size = 0; |
|
return ret; |
|
} |
|
|
|
static int videotoolbox_hevc_start_frame(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
return 0; |
|
} |
|
|
|
static int videotoolbox_hevc_decode_slice(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
return videotoolbox_common_decode_slice(avctx, buffer, size); |
|
} |
|
|
|
|
|
static int videotoolbox_hevc_decode_params(AVCodecContext *avctx, |
|
int type, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
return videotoolbox_common_decode_slice(avctx, buffer, size); |
|
} |
|
|
|
static int videotoolbox_hevc_end_frame(AVCodecContext *avctx) |
|
{ |
|
HEVCContext *h = avctx->priv_data; |
|
AVFrame *frame = h->ref->frame; |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
h->output_frame->crop_right = 0; |
|
h->output_frame->crop_left = 0; |
|
h->output_frame->crop_top = 0; |
|
h->output_frame->crop_bottom = 0; |
|
|
|
int ret = ff_videotoolbox_common_end_frame(avctx, frame); |
|
vtctx->bitstream_size = 0; |
|
return ret; |
|
} |
|
|
|
static int videotoolbox_mpeg_start_frame(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
return ff_videotoolbox_buffer_copy(vtctx, buffer, size); |
|
} |
|
|
|
static int videotoolbox_mpeg_decode_slice(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
return 0; |
|
} |
|
|
|
static int videotoolbox_mpeg_end_frame(AVCodecContext *avctx) |
|
{ |
|
MpegEncContext *s = avctx->priv_data; |
|
AVFrame *frame = s->current_picture_ptr->f; |
|
|
|
return ff_videotoolbox_common_end_frame(avctx, frame); |
|
} |
|
|
|
static int videotoolbox_prores_start_frame(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
return 0; |
|
} |
|
|
|
static int videotoolbox_prores_decode_slice(AVCodecContext *avctx, |
|
const uint8_t *buffer, |
|
uint32_t size) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
|
|
return ff_videotoolbox_buffer_copy(vtctx, buffer, size); |
|
} |
|
|
|
static int videotoolbox_prores_end_frame(AVCodecContext *avctx) |
|
{ |
|
ProresContext *ctx = avctx->priv_data; |
|
AVFrame *frame = ctx->frame; |
|
|
|
return ff_videotoolbox_common_end_frame(avctx, frame); |
|
} |
|
|
|
static enum AVPixelFormat videotoolbox_best_pixel_format(AVCodecContext *avctx) { |
|
const AVPixFmtDescriptor *descriptor = av_pix_fmt_desc_get(avctx->sw_pix_fmt); |
|
if (!descriptor) |
|
return AV_PIX_FMT_NV12; // same as av_videotoolbox_alloc_context() |
|
|
|
int depth = descriptor->comp[0].depth; |
|
|
|
if (descriptor->flags & AV_PIX_FMT_FLAG_ALPHA) |
|
return AV_PIX_FMT_AYUV64; |
|
|
|
#if HAVE_KCVPIXELFORMATTYPE_444YPCBCR16BIPLANARVIDEORANGE |
|
if (depth > 10) |
|
return descriptor->log2_chroma_w == 0 ? AV_PIX_FMT_P416 : AV_PIX_FMT_P216; |
|
#endif |
|
|
|
#if HAVE_KCVPIXELFORMATTYPE_444YPCBCR10BIPLANARVIDEORANGE |
|
if (descriptor->log2_chroma_w == 0) { |
|
#if HAVE_KCVPIXELFORMATTYPE_444YPCBCR8BIPLANARVIDEORANGE |
|
if (depth <= 8) |
|
return AV_PIX_FMT_NV24; |
|
#endif |
|
return AV_PIX_FMT_P410; |
|
} |
|
#endif |
|
#if HAVE_KCVPIXELFORMATTYPE_422YPCBCR10BIPLANARVIDEORANGE |
|
if (descriptor->log2_chroma_h == 0) { |
|
#if HAVE_KCVPIXELFORMATTYPE_422YPCBCR8BIPLANARVIDEORANGE |
|
if (depth <= 8) |
|
return AV_PIX_FMT_NV16; |
|
#endif |
|
return AV_PIX_FMT_P210; |
|
} |
|
#endif |
|
#if HAVE_KCVPIXELFORMATTYPE_420YPCBCR10BIPLANARVIDEORANGE |
|
if (depth > 8) { |
|
return AV_PIX_FMT_P010; |
|
} |
|
#endif |
|
|
|
return AV_PIX_FMT_NV12; |
|
} |
|
|
|
int ff_videotoolbox_common_init(AVCodecContext *avctx) |
|
{ |
|
VTContext *vtctx = avctx->internal->hwaccel_priv_data; |
|
AVHWFramesContext *hw_frames; |
|
int err; |
|
|
|
// Old API - do nothing. |
|
if (avctx->hwaccel_context) |
|
return 0; |
|
|
|
if (!avctx->hw_frames_ctx && !avctx->hw_device_ctx) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Either hw_frames_ctx or hw_device_ctx must be set.\n"); |
|
return AVERROR(EINVAL); |
|
} |
|
|
|
vtctx->vt_ctx = av_videotoolbox_alloc_context(); |
|
if (!vtctx->vt_ctx) { |
|
err = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
if (avctx->hw_frames_ctx) { |
|
hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data; |
|
} else { |
|
avctx->hw_frames_ctx = av_hwframe_ctx_alloc(avctx->hw_device_ctx); |
|
if (!avctx->hw_frames_ctx) { |
|
err = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data; |
|
hw_frames->format = AV_PIX_FMT_VIDEOTOOLBOX; |
|
hw_frames->sw_format = videotoolbox_best_pixel_format(avctx); |
|
hw_frames->width = avctx->width; |
|
hw_frames->height = avctx->height; |
|
|
|
err = av_hwframe_ctx_init(avctx->hw_frames_ctx); |
|
if (err < 0) { |
|
av_buffer_unref(&avctx->hw_frames_ctx); |
|
goto fail; |
|
} |
|
} |
|
|
|
vtctx->cached_hw_frames_ctx = av_buffer_ref(avctx->hw_frames_ctx); |
|
if (!vtctx->cached_hw_frames_ctx) { |
|
err = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
bool full_range = avctx->color_range == AVCOL_RANGE_JPEG; |
|
vtctx->vt_ctx->cv_pix_fmt_type = |
|
av_map_videotoolbox_format_from_pixfmt2(hw_frames->sw_format, full_range); |
|
if (!vtctx->vt_ctx->cv_pix_fmt_type) { |
|
const AVPixFmtDescriptor *attempted_format = |
|
av_pix_fmt_desc_get(hw_frames->sw_format); |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Failed to map underlying FFmpeg pixel format %s (%s range) to " |
|
"a VideoToolbox format!\n", |
|
attempted_format ? attempted_format->name : "<unknown>", |
|
av_color_range_name(avctx->color_range)); |
|
err = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
err = videotoolbox_start(avctx); |
|
if (err < 0) |
|
goto fail; |
|
|
|
return 0; |
|
|
|
fail: |
|
ff_videotoolbox_uninit(avctx); |
|
return err; |
|
} |
|
|
|
int ff_videotoolbox_frame_params(AVCodecContext *avctx, |
|
AVBufferRef *hw_frames_ctx) |
|
{ |
|
AVHWFramesContext *frames_ctx = (AVHWFramesContext*)hw_frames_ctx->data; |
|
|
|
frames_ctx->format = AV_PIX_FMT_VIDEOTOOLBOX; |
|
frames_ctx->width = avctx->coded_width; |
|
frames_ctx->height = avctx->coded_height; |
|
frames_ctx->sw_format = videotoolbox_best_pixel_format(avctx); |
|
|
|
return 0; |
|
} |
|
|
|
const AVHWAccel ff_h263_videotoolbox_hwaccel = { |
|
.name = "h263_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_H263, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_mpeg_start_frame, |
|
.decode_slice = videotoolbox_mpeg_decode_slice, |
|
.end_frame = videotoolbox_mpeg_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_hevc_videotoolbox_hwaccel = { |
|
.name = "hevc_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_HEVC, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_hevc_start_frame, |
|
.decode_slice = videotoolbox_hevc_decode_slice, |
|
.decode_params = videotoolbox_hevc_decode_params, |
|
.end_frame = videotoolbox_hevc_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_h264_videotoolbox_hwaccel = { |
|
.name = "h264_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_H264, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = ff_videotoolbox_h264_start_frame, |
|
.decode_slice = ff_videotoolbox_h264_decode_slice, |
|
.decode_params = videotoolbox_h264_decode_params, |
|
.end_frame = videotoolbox_h264_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_mpeg1_videotoolbox_hwaccel = { |
|
.name = "mpeg1_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_MPEG1VIDEO, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_mpeg_start_frame, |
|
.decode_slice = videotoolbox_mpeg_decode_slice, |
|
.end_frame = videotoolbox_mpeg_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_mpeg2_videotoolbox_hwaccel = { |
|
.name = "mpeg2_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_MPEG2VIDEO, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_mpeg_start_frame, |
|
.decode_slice = videotoolbox_mpeg_decode_slice, |
|
.end_frame = videotoolbox_mpeg_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_mpeg4_videotoolbox_hwaccel = { |
|
.name = "mpeg4_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_MPEG4, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_mpeg_start_frame, |
|
.decode_slice = videotoolbox_mpeg_decode_slice, |
|
.end_frame = videotoolbox_mpeg_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
const AVHWAccel ff_prores_videotoolbox_hwaccel = { |
|
.name = "prores_videotoolbox", |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_PRORES, |
|
.pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX, |
|
.alloc_frame = ff_videotoolbox_alloc_frame, |
|
.start_frame = videotoolbox_prores_start_frame, |
|
.decode_slice = videotoolbox_prores_decode_slice, |
|
.end_frame = videotoolbox_prores_end_frame, |
|
.frame_params = ff_videotoolbox_frame_params, |
|
.init = ff_videotoolbox_common_init, |
|
.uninit = ff_videotoolbox_uninit, |
|
.priv_data_size = sizeof(VTContext), |
|
}; |
|
|
|
static AVVideotoolboxContext *av_videotoolbox_alloc_context_with_pix_fmt(enum AVPixelFormat pix_fmt, |
|
bool full_range) |
|
{ |
|
AVVideotoolboxContext *ret = av_mallocz(sizeof(*ret)); |
|
|
|
if (ret) { |
|
ret->output_callback = videotoolbox_decoder_callback; |
|
|
|
OSType cv_pix_fmt_type = av_map_videotoolbox_format_from_pixfmt2(pix_fmt, full_range); |
|
if (cv_pix_fmt_type == 0) { |
|
cv_pix_fmt_type = kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange; |
|
} |
|
ret->cv_pix_fmt_type = cv_pix_fmt_type; |
|
} |
|
|
|
return ret; |
|
} |
|
|
|
AVVideotoolboxContext *av_videotoolbox_alloc_context(void) |
|
{ |
|
return av_videotoolbox_alloc_context_with_pix_fmt(AV_PIX_FMT_NONE, false); |
|
} |
|
|
|
int av_videotoolbox_default_init(AVCodecContext *avctx) |
|
{ |
|
return av_videotoolbox_default_init2(avctx, NULL); |
|
} |
|
|
|
int av_videotoolbox_default_init2(AVCodecContext *avctx, AVVideotoolboxContext *vtctx) |
|
{ |
|
enum AVPixelFormat pix_fmt = videotoolbox_best_pixel_format(avctx); |
|
bool full_range = avctx->color_range == AVCOL_RANGE_JPEG; |
|
avctx->hwaccel_context = vtctx ?: av_videotoolbox_alloc_context_with_pix_fmt(pix_fmt, full_range); |
|
if (!avctx->hwaccel_context) |
|
return AVERROR(ENOMEM); |
|
return videotoolbox_start(avctx); |
|
} |
|
|
|
void av_videotoolbox_default_free(AVCodecContext *avctx) |
|
{ |
|
|
|
videotoolbox_stop(avctx); |
|
av_freep(&avctx->hwaccel_context); |
|
} |
|
#endif /* CONFIG_VIDEOTOOLBOX */
|
|
|