mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1129 lines
35 KiB
1129 lines
35 KiB
/* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include "channel_layout.h" |
|
#include "avassert.h" |
|
#include "buffer.h" |
|
#include "common.h" |
|
#include "cpu.h" |
|
#include "dict.h" |
|
#include "frame.h" |
|
#include "imgutils.h" |
|
#include "mem.h" |
|
#include "samplefmt.h" |
|
#include "hwcontext.h" |
|
|
|
static const AVSideDataDescriptor sd_props[] = { |
|
[AV_FRAME_DATA_PANSCAN] = { "AVPanScan" }, |
|
[AV_FRAME_DATA_A53_CC] = { "ATSC A53 Part 4 Closed Captions" }, |
|
[AV_FRAME_DATA_MATRIXENCODING] = { "AVMatrixEncoding" }, |
|
[AV_FRAME_DATA_DOWNMIX_INFO] = { "Metadata relevant to a downmix procedure" }, |
|
[AV_FRAME_DATA_AFD] = { "Active format description" }, |
|
[AV_FRAME_DATA_MOTION_VECTORS] = { "Motion vectors" }, |
|
[AV_FRAME_DATA_SKIP_SAMPLES] = { "Skip samples" }, |
|
[AV_FRAME_DATA_GOP_TIMECODE] = { "GOP timecode" }, |
|
[AV_FRAME_DATA_S12M_TIMECODE] = { "SMPTE 12-1 timecode" }, |
|
[AV_FRAME_DATA_DYNAMIC_HDR_PLUS] = { "HDR Dynamic Metadata SMPTE2094-40 (HDR10+)" }, |
|
[AV_FRAME_DATA_DYNAMIC_HDR_VIVID] = { "HDR Dynamic Metadata CUVA 005.1 2021 (Vivid)" }, |
|
[AV_FRAME_DATA_REGIONS_OF_INTEREST] = { "Regions Of Interest" }, |
|
[AV_FRAME_DATA_VIDEO_ENC_PARAMS] = { "Video encoding parameters" }, |
|
[AV_FRAME_DATA_FILM_GRAIN_PARAMS] = { "Film grain parameters" }, |
|
[AV_FRAME_DATA_DETECTION_BBOXES] = { "Bounding boxes for object detection and classification" }, |
|
[AV_FRAME_DATA_DOVI_RPU_BUFFER] = { "Dolby Vision RPU Data" }, |
|
[AV_FRAME_DATA_DOVI_METADATA] = { "Dolby Vision Metadata" }, |
|
[AV_FRAME_DATA_STEREO3D] = { "Stereo 3D", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_REPLAYGAIN] = { "AVReplayGain", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_DISPLAYMATRIX] = { "3x3 displaymatrix", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_AUDIO_SERVICE_TYPE] = { "Audio service type", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_MASTERING_DISPLAY_METADATA] = { "Mastering display metadata", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_CONTENT_LIGHT_LEVEL] = { "Content light level metadata", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_AMBIENT_VIEWING_ENVIRONMENT] = { "Ambient viewing environment", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_SPHERICAL] = { "Spherical Mapping", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_ICC_PROFILE] = { "ICC profile", AV_SIDE_DATA_PROP_GLOBAL }, |
|
[AV_FRAME_DATA_SEI_UNREGISTERED] = { "H.26[45] User Data Unregistered SEI message", AV_SIDE_DATA_PROP_MULTI }, |
|
}; |
|
|
|
static void get_frame_defaults(AVFrame *frame) |
|
{ |
|
memset(frame, 0, sizeof(*frame)); |
|
|
|
frame->pts = |
|
frame->pkt_dts = AV_NOPTS_VALUE; |
|
frame->best_effort_timestamp = AV_NOPTS_VALUE; |
|
frame->duration = 0; |
|
#if FF_API_FRAME_PKT |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
frame->pkt_pos = -1; |
|
frame->pkt_size = -1; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
frame->time_base = (AVRational){ 0, 1 }; |
|
frame->sample_aspect_ratio = (AVRational){ 0, 1 }; |
|
frame->format = -1; /* unknown */ |
|
frame->extended_data = frame->data; |
|
frame->color_primaries = AVCOL_PRI_UNSPECIFIED; |
|
frame->color_trc = AVCOL_TRC_UNSPECIFIED; |
|
frame->colorspace = AVCOL_SPC_UNSPECIFIED; |
|
frame->color_range = AVCOL_RANGE_UNSPECIFIED; |
|
frame->chroma_location = AVCHROMA_LOC_UNSPECIFIED; |
|
frame->flags = 0; |
|
} |
|
|
|
static void free_side_data(AVFrameSideData **ptr_sd) |
|
{ |
|
AVFrameSideData *sd = *ptr_sd; |
|
|
|
av_buffer_unref(&sd->buf); |
|
av_dict_free(&sd->metadata); |
|
av_freep(ptr_sd); |
|
} |
|
|
|
static void wipe_side_data(AVFrameSideData ***sd, int *nb_side_data) |
|
{ |
|
for (int i = 0; i < *nb_side_data; i++) { |
|
free_side_data(&((*sd)[i])); |
|
} |
|
*nb_side_data = 0; |
|
|
|
av_freep(sd); |
|
} |
|
|
|
static void frame_side_data_wipe(AVFrame *frame) |
|
{ |
|
wipe_side_data(&frame->side_data, &frame->nb_side_data); |
|
} |
|
|
|
void av_frame_side_data_free(AVFrameSideData ***sd, int *nb_sd) |
|
{ |
|
wipe_side_data(sd, nb_sd); |
|
} |
|
|
|
static void remove_side_data(AVFrameSideData ***sd, int *nb_side_data, |
|
const enum AVFrameSideDataType type) |
|
{ |
|
for (int i = *nb_side_data - 1; i >= 0; i--) { |
|
AVFrameSideData *entry = ((*sd)[i]); |
|
if (entry->type != type) |
|
continue; |
|
|
|
free_side_data(&entry); |
|
|
|
((*sd)[i]) = ((*sd)[*nb_side_data - 1]); |
|
(*nb_side_data)--; |
|
} |
|
} |
|
|
|
static void remove_side_data_by_entry(AVFrameSideData ***sd, int *nb_sd, |
|
const AVFrameSideData *target) |
|
{ |
|
for (int i = *nb_sd - 1; i >= 0; i--) { |
|
AVFrameSideData *entry = ((*sd)[i]); |
|
if (entry != target) |
|
continue; |
|
|
|
free_side_data(&entry); |
|
|
|
((*sd)[i]) = ((*sd)[*nb_sd - 1]); |
|
(*nb_sd)--; |
|
|
|
return; |
|
} |
|
} |
|
|
|
AVFrame *av_frame_alloc(void) |
|
{ |
|
AVFrame *frame = av_malloc(sizeof(*frame)); |
|
|
|
if (!frame) |
|
return NULL; |
|
|
|
get_frame_defaults(frame); |
|
|
|
return frame; |
|
} |
|
|
|
void av_frame_free(AVFrame **frame) |
|
{ |
|
if (!frame || !*frame) |
|
return; |
|
|
|
av_frame_unref(*frame); |
|
av_freep(frame); |
|
} |
|
|
|
#define ALIGN (HAVE_SIMD_ALIGN_64 ? 64 : 32) |
|
|
|
static int get_video_buffer(AVFrame *frame, int align) |
|
{ |
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format); |
|
int ret, padded_height, total_size; |
|
int plane_padding = FFMAX(16 + 16/*STRIDE_ALIGN*/, align); |
|
ptrdiff_t linesizes[4]; |
|
size_t sizes[4]; |
|
|
|
if (!desc) |
|
return AVERROR(EINVAL); |
|
|
|
if ((ret = av_image_check_size(frame->width, frame->height, 0, NULL)) < 0) |
|
return ret; |
|
|
|
if (!frame->linesize[0]) { |
|
if (align <= 0) |
|
align = ALIGN; |
|
|
|
for (int i = 1; i <= align; i += i) { |
|
ret = av_image_fill_linesizes(frame->linesize, frame->format, |
|
FFALIGN(frame->width, i)); |
|
if (ret < 0) |
|
return ret; |
|
if (!(frame->linesize[0] & (align-1))) |
|
break; |
|
} |
|
|
|
for (int i = 0; i < 4 && frame->linesize[i]; i++) |
|
frame->linesize[i] = FFALIGN(frame->linesize[i], align); |
|
} |
|
|
|
for (int i = 0; i < 4; i++) |
|
linesizes[i] = frame->linesize[i]; |
|
|
|
padded_height = FFALIGN(frame->height, 32); |
|
if ((ret = av_image_fill_plane_sizes(sizes, frame->format, |
|
padded_height, linesizes)) < 0) |
|
return ret; |
|
|
|
total_size = 4*plane_padding; |
|
for (int i = 0; i < 4; i++) { |
|
if (sizes[i] > INT_MAX - total_size) |
|
return AVERROR(EINVAL); |
|
total_size += sizes[i]; |
|
} |
|
|
|
frame->buf[0] = av_buffer_alloc(total_size); |
|
if (!frame->buf[0]) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
|
|
if ((ret = av_image_fill_pointers(frame->data, frame->format, padded_height, |
|
frame->buf[0]->data, frame->linesize)) < 0) |
|
goto fail; |
|
|
|
for (int i = 1; i < 4; i++) { |
|
if (frame->data[i]) |
|
frame->data[i] += i * plane_padding; |
|
} |
|
|
|
frame->extended_data = frame->data; |
|
|
|
return 0; |
|
fail: |
|
av_frame_unref(frame); |
|
return ret; |
|
} |
|
|
|
static int get_audio_buffer(AVFrame *frame, int align) |
|
{ |
|
int planar = av_sample_fmt_is_planar(frame->format); |
|
int channels, planes; |
|
int ret; |
|
|
|
channels = frame->ch_layout.nb_channels; |
|
planes = planar ? channels : 1; |
|
if (!frame->linesize[0]) { |
|
ret = av_samples_get_buffer_size(&frame->linesize[0], channels, |
|
frame->nb_samples, frame->format, |
|
align); |
|
if (ret < 0) |
|
return ret; |
|
} |
|
|
|
if (planes > AV_NUM_DATA_POINTERS) { |
|
frame->extended_data = av_calloc(planes, |
|
sizeof(*frame->extended_data)); |
|
frame->extended_buf = av_calloc(planes - AV_NUM_DATA_POINTERS, |
|
sizeof(*frame->extended_buf)); |
|
if (!frame->extended_data || !frame->extended_buf) { |
|
av_freep(&frame->extended_data); |
|
av_freep(&frame->extended_buf); |
|
return AVERROR(ENOMEM); |
|
} |
|
frame->nb_extended_buf = planes - AV_NUM_DATA_POINTERS; |
|
} else |
|
frame->extended_data = frame->data; |
|
|
|
for (int i = 0; i < FFMIN(planes, AV_NUM_DATA_POINTERS); i++) { |
|
frame->buf[i] = av_buffer_alloc(frame->linesize[0]); |
|
if (!frame->buf[i]) { |
|
av_frame_unref(frame); |
|
return AVERROR(ENOMEM); |
|
} |
|
frame->extended_data[i] = frame->data[i] = frame->buf[i]->data; |
|
} |
|
for (int i = 0; i < planes - AV_NUM_DATA_POINTERS; i++) { |
|
frame->extended_buf[i] = av_buffer_alloc(frame->linesize[0]); |
|
if (!frame->extended_buf[i]) { |
|
av_frame_unref(frame); |
|
return AVERROR(ENOMEM); |
|
} |
|
frame->extended_data[i + AV_NUM_DATA_POINTERS] = frame->extended_buf[i]->data; |
|
} |
|
return 0; |
|
|
|
} |
|
|
|
int av_frame_get_buffer(AVFrame *frame, int align) |
|
{ |
|
if (frame->format < 0) |
|
return AVERROR(EINVAL); |
|
|
|
if (frame->width > 0 && frame->height > 0) |
|
return get_video_buffer(frame, align); |
|
else if (frame->nb_samples > 0 && |
|
(av_channel_layout_check(&frame->ch_layout))) |
|
return get_audio_buffer(frame, align); |
|
|
|
return AVERROR(EINVAL); |
|
} |
|
|
|
static int frame_copy_props(AVFrame *dst, const AVFrame *src, int force_copy) |
|
{ |
|
int ret; |
|
|
|
#if FF_API_FRAME_KEY |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
dst->key_frame = src->key_frame; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
dst->pict_type = src->pict_type; |
|
dst->sample_aspect_ratio = src->sample_aspect_ratio; |
|
dst->crop_top = src->crop_top; |
|
dst->crop_bottom = src->crop_bottom; |
|
dst->crop_left = src->crop_left; |
|
dst->crop_right = src->crop_right; |
|
dst->pts = src->pts; |
|
dst->duration = src->duration; |
|
dst->repeat_pict = src->repeat_pict; |
|
#if FF_API_INTERLACED_FRAME |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
dst->interlaced_frame = src->interlaced_frame; |
|
dst->top_field_first = src->top_field_first; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
#if FF_API_PALETTE_HAS_CHANGED |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
dst->palette_has_changed = src->palette_has_changed; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
dst->sample_rate = src->sample_rate; |
|
dst->opaque = src->opaque; |
|
dst->pkt_dts = src->pkt_dts; |
|
#if FF_API_FRAME_PKT |
|
FF_DISABLE_DEPRECATION_WARNINGS |
|
dst->pkt_pos = src->pkt_pos; |
|
dst->pkt_size = src->pkt_size; |
|
FF_ENABLE_DEPRECATION_WARNINGS |
|
#endif |
|
dst->time_base = src->time_base; |
|
dst->quality = src->quality; |
|
dst->best_effort_timestamp = src->best_effort_timestamp; |
|
dst->flags = src->flags; |
|
dst->decode_error_flags = src->decode_error_flags; |
|
dst->color_primaries = src->color_primaries; |
|
dst->color_trc = src->color_trc; |
|
dst->colorspace = src->colorspace; |
|
dst->color_range = src->color_range; |
|
dst->chroma_location = src->chroma_location; |
|
|
|
av_dict_copy(&dst->metadata, src->metadata, 0); |
|
|
|
for (int i = 0; i < src->nb_side_data; i++) { |
|
const AVFrameSideData *sd_src = src->side_data[i]; |
|
AVFrameSideData *sd_dst; |
|
if ( sd_src->type == AV_FRAME_DATA_PANSCAN |
|
&& (src->width != dst->width || src->height != dst->height)) |
|
continue; |
|
if (force_copy) { |
|
sd_dst = av_frame_new_side_data(dst, sd_src->type, |
|
sd_src->size); |
|
if (!sd_dst) { |
|
frame_side_data_wipe(dst); |
|
return AVERROR(ENOMEM); |
|
} |
|
memcpy(sd_dst->data, sd_src->data, sd_src->size); |
|
} else { |
|
AVBufferRef *ref = av_buffer_ref(sd_src->buf); |
|
sd_dst = av_frame_new_side_data_from_buf(dst, sd_src->type, ref); |
|
if (!sd_dst) { |
|
av_buffer_unref(&ref); |
|
frame_side_data_wipe(dst); |
|
return AVERROR(ENOMEM); |
|
} |
|
} |
|
av_dict_copy(&sd_dst->metadata, sd_src->metadata, 0); |
|
} |
|
|
|
ret = av_buffer_replace(&dst->opaque_ref, src->opaque_ref); |
|
ret |= av_buffer_replace(&dst->private_ref, src->private_ref); |
|
return ret; |
|
} |
|
|
|
int av_frame_ref(AVFrame *dst, const AVFrame *src) |
|
{ |
|
int ret = 0; |
|
|
|
av_assert1(dst->width == 0 && dst->height == 0); |
|
av_assert1(dst->ch_layout.nb_channels == 0 && |
|
dst->ch_layout.order == AV_CHANNEL_ORDER_UNSPEC); |
|
|
|
dst->format = src->format; |
|
dst->width = src->width; |
|
dst->height = src->height; |
|
dst->nb_samples = src->nb_samples; |
|
|
|
ret = frame_copy_props(dst, src, 0); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
ret = av_channel_layout_copy(&dst->ch_layout, &src->ch_layout); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
/* duplicate the frame data if it's not refcounted */ |
|
if (!src->buf[0]) { |
|
ret = av_frame_get_buffer(dst, 0); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
ret = av_frame_copy(dst, src); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
return 0; |
|
} |
|
|
|
/* ref the buffers */ |
|
for (int i = 0; i < FF_ARRAY_ELEMS(src->buf); i++) { |
|
if (!src->buf[i]) |
|
continue; |
|
dst->buf[i] = av_buffer_ref(src->buf[i]); |
|
if (!dst->buf[i]) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
} |
|
|
|
if (src->extended_buf) { |
|
dst->extended_buf = av_calloc(src->nb_extended_buf, |
|
sizeof(*dst->extended_buf)); |
|
if (!dst->extended_buf) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
dst->nb_extended_buf = src->nb_extended_buf; |
|
|
|
for (int i = 0; i < src->nb_extended_buf; i++) { |
|
dst->extended_buf[i] = av_buffer_ref(src->extended_buf[i]); |
|
if (!dst->extended_buf[i]) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
} |
|
} |
|
|
|
if (src->hw_frames_ctx) { |
|
dst->hw_frames_ctx = av_buffer_ref(src->hw_frames_ctx); |
|
if (!dst->hw_frames_ctx) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
} |
|
|
|
/* duplicate extended data */ |
|
if (src->extended_data != src->data) { |
|
int ch = dst->ch_layout.nb_channels; |
|
|
|
if (!ch) { |
|
ret = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
dst->extended_data = av_malloc_array(sizeof(*dst->extended_data), ch); |
|
if (!dst->extended_data) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
memcpy(dst->extended_data, src->extended_data, sizeof(*src->extended_data) * ch); |
|
} else |
|
dst->extended_data = dst->data; |
|
|
|
memcpy(dst->data, src->data, sizeof(src->data)); |
|
memcpy(dst->linesize, src->linesize, sizeof(src->linesize)); |
|
|
|
return 0; |
|
|
|
fail: |
|
av_frame_unref(dst); |
|
return ret; |
|
} |
|
|
|
int av_frame_replace(AVFrame *dst, const AVFrame *src) |
|
{ |
|
int ret = 0; |
|
|
|
if (dst == src) |
|
return AVERROR(EINVAL); |
|
|
|
if (!src->buf[0]) { |
|
av_frame_unref(dst); |
|
|
|
/* duplicate the frame data if it's not refcounted */ |
|
if ( src->data[0] || src->data[1] |
|
|| src->data[2] || src->data[3]) |
|
return av_frame_ref(dst, src); |
|
|
|
ret = frame_copy_props(dst, src, 0); |
|
if (ret < 0) |
|
goto fail; |
|
} |
|
|
|
dst->format = src->format; |
|
dst->width = src->width; |
|
dst->height = src->height; |
|
dst->nb_samples = src->nb_samples; |
|
|
|
ret = av_channel_layout_copy(&dst->ch_layout, &src->ch_layout); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
frame_side_data_wipe(dst); |
|
av_dict_free(&dst->metadata); |
|
ret = frame_copy_props(dst, src, 0); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
/* replace the buffers */ |
|
for (int i = 0; i < FF_ARRAY_ELEMS(src->buf); i++) { |
|
ret = av_buffer_replace(&dst->buf[i], src->buf[i]); |
|
if (ret < 0) |
|
goto fail; |
|
} |
|
|
|
if (src->extended_buf) { |
|
if (dst->nb_extended_buf != src->nb_extended_buf) { |
|
int nb_extended_buf = FFMIN(dst->nb_extended_buf, src->nb_extended_buf); |
|
void *tmp; |
|
|
|
for (int i = nb_extended_buf; i < dst->nb_extended_buf; i++) |
|
av_buffer_unref(&dst->extended_buf[i]); |
|
|
|
tmp = av_realloc_array(dst->extended_buf, sizeof(*dst->extended_buf), |
|
src->nb_extended_buf); |
|
if (!tmp) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
dst->extended_buf = tmp; |
|
dst->nb_extended_buf = src->nb_extended_buf; |
|
|
|
memset(&dst->extended_buf[nb_extended_buf], 0, |
|
(src->nb_extended_buf - nb_extended_buf) * sizeof(*dst->extended_buf)); |
|
} |
|
|
|
for (int i = 0; i < src->nb_extended_buf; i++) { |
|
ret = av_buffer_replace(&dst->extended_buf[i], src->extended_buf[i]); |
|
if (ret < 0) |
|
goto fail; |
|
} |
|
} else if (dst->extended_buf) { |
|
for (int i = 0; i < dst->nb_extended_buf; i++) |
|
av_buffer_unref(&dst->extended_buf[i]); |
|
av_freep(&dst->extended_buf); |
|
} |
|
|
|
ret = av_buffer_replace(&dst->hw_frames_ctx, src->hw_frames_ctx); |
|
if (ret < 0) |
|
goto fail; |
|
|
|
if (dst->extended_data != dst->data) |
|
av_freep(&dst->extended_data); |
|
|
|
if (src->extended_data != src->data) { |
|
int ch = dst->ch_layout.nb_channels; |
|
|
|
if (!ch) { |
|
ret = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
if (ch > SIZE_MAX / sizeof(*dst->extended_data)) |
|
goto fail; |
|
|
|
dst->extended_data = av_memdup(src->extended_data, sizeof(*dst->extended_data) * ch); |
|
if (!dst->extended_data) { |
|
ret = AVERROR(ENOMEM); |
|
goto fail; |
|
} |
|
} else |
|
dst->extended_data = dst->data; |
|
|
|
memcpy(dst->data, src->data, sizeof(src->data)); |
|
memcpy(dst->linesize, src->linesize, sizeof(src->linesize)); |
|
|
|
return 0; |
|
|
|
fail: |
|
av_frame_unref(dst); |
|
return ret; |
|
} |
|
|
|
AVFrame *av_frame_clone(const AVFrame *src) |
|
{ |
|
AVFrame *ret = av_frame_alloc(); |
|
|
|
if (!ret) |
|
return NULL; |
|
|
|
if (av_frame_ref(ret, src) < 0) |
|
av_frame_free(&ret); |
|
|
|
return ret; |
|
} |
|
|
|
void av_frame_unref(AVFrame *frame) |
|
{ |
|
if (!frame) |
|
return; |
|
|
|
frame_side_data_wipe(frame); |
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++) |
|
av_buffer_unref(&frame->buf[i]); |
|
for (int i = 0; i < frame->nb_extended_buf; i++) |
|
av_buffer_unref(&frame->extended_buf[i]); |
|
av_freep(&frame->extended_buf); |
|
av_dict_free(&frame->metadata); |
|
|
|
av_buffer_unref(&frame->hw_frames_ctx); |
|
|
|
av_buffer_unref(&frame->opaque_ref); |
|
av_buffer_unref(&frame->private_ref); |
|
|
|
if (frame->extended_data != frame->data) |
|
av_freep(&frame->extended_data); |
|
|
|
av_channel_layout_uninit(&frame->ch_layout); |
|
|
|
get_frame_defaults(frame); |
|
} |
|
|
|
void av_frame_move_ref(AVFrame *dst, AVFrame *src) |
|
{ |
|
av_assert1(dst->width == 0 && dst->height == 0); |
|
av_assert1(dst->ch_layout.nb_channels == 0 && |
|
dst->ch_layout.order == AV_CHANNEL_ORDER_UNSPEC); |
|
|
|
*dst = *src; |
|
if (src->extended_data == src->data) |
|
dst->extended_data = dst->data; |
|
get_frame_defaults(src); |
|
} |
|
|
|
int av_frame_is_writable(AVFrame *frame) |
|
{ |
|
int ret = 1; |
|
|
|
/* assume non-refcounted frames are not writable */ |
|
if (!frame->buf[0]) |
|
return 0; |
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++) |
|
if (frame->buf[i]) |
|
ret &= !!av_buffer_is_writable(frame->buf[i]); |
|
for (int i = 0; i < frame->nb_extended_buf; i++) |
|
ret &= !!av_buffer_is_writable(frame->extended_buf[i]); |
|
|
|
return ret; |
|
} |
|
|
|
int av_frame_make_writable(AVFrame *frame) |
|
{ |
|
AVFrame tmp; |
|
int ret; |
|
|
|
if (av_frame_is_writable(frame)) |
|
return 0; |
|
|
|
memset(&tmp, 0, sizeof(tmp)); |
|
tmp.format = frame->format; |
|
tmp.width = frame->width; |
|
tmp.height = frame->height; |
|
tmp.nb_samples = frame->nb_samples; |
|
ret = av_channel_layout_copy(&tmp.ch_layout, &frame->ch_layout); |
|
if (ret < 0) { |
|
av_frame_unref(&tmp); |
|
return ret; |
|
} |
|
|
|
if (frame->hw_frames_ctx) |
|
ret = av_hwframe_get_buffer(frame->hw_frames_ctx, &tmp, 0); |
|
else |
|
ret = av_frame_get_buffer(&tmp, 0); |
|
if (ret < 0) |
|
return ret; |
|
|
|
ret = av_frame_copy(&tmp, frame); |
|
if (ret < 0) { |
|
av_frame_unref(&tmp); |
|
return ret; |
|
} |
|
|
|
ret = av_frame_copy_props(&tmp, frame); |
|
if (ret < 0) { |
|
av_frame_unref(&tmp); |
|
return ret; |
|
} |
|
|
|
av_frame_unref(frame); |
|
|
|
*frame = tmp; |
|
if (tmp.data == tmp.extended_data) |
|
frame->extended_data = frame->data; |
|
|
|
return 0; |
|
} |
|
|
|
int av_frame_copy_props(AVFrame *dst, const AVFrame *src) |
|
{ |
|
return frame_copy_props(dst, src, 1); |
|
} |
|
|
|
AVBufferRef *av_frame_get_plane_buffer(const AVFrame *frame, int plane) |
|
{ |
|
uint8_t *data; |
|
int planes; |
|
|
|
if (frame->nb_samples) { |
|
int channels = frame->ch_layout.nb_channels; |
|
if (!channels) |
|
return NULL; |
|
planes = av_sample_fmt_is_planar(frame->format) ? channels : 1; |
|
} else |
|
planes = 4; |
|
|
|
if (plane < 0 || plane >= planes || !frame->extended_data[plane]) |
|
return NULL; |
|
data = frame->extended_data[plane]; |
|
|
|
for (int i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++) { |
|
AVBufferRef *buf = frame->buf[i]; |
|
if (data >= buf->data && data < buf->data + buf->size) |
|
return buf; |
|
} |
|
for (int i = 0; i < frame->nb_extended_buf; i++) { |
|
AVBufferRef *buf = frame->extended_buf[i]; |
|
if (data >= buf->data && data < buf->data + buf->size) |
|
return buf; |
|
} |
|
return NULL; |
|
} |
|
|
|
static AVFrameSideData *add_side_data_from_buf_ext(AVFrameSideData ***sd, |
|
int *nb_sd, |
|
enum AVFrameSideDataType type, |
|
AVBufferRef *buf, uint8_t *data, |
|
size_t size) |
|
{ |
|
AVFrameSideData *ret, **tmp; |
|
|
|
// *nb_sd + 1 needs to fit into an int and a size_t. |
|
if ((unsigned)*nb_sd >= FFMIN(INT_MAX, SIZE_MAX)) |
|
return NULL; |
|
|
|
tmp = av_realloc_array(*sd, sizeof(**sd), *nb_sd + 1); |
|
if (!tmp) |
|
return NULL; |
|
*sd = tmp; |
|
|
|
ret = av_mallocz(sizeof(*ret)); |
|
if (!ret) |
|
return NULL; |
|
|
|
ret->buf = buf; |
|
ret->data = data; |
|
ret->size = size; |
|
ret->type = type; |
|
|
|
(*sd)[(*nb_sd)++] = ret; |
|
|
|
return ret; |
|
} |
|
|
|
static AVFrameSideData *add_side_data_from_buf(AVFrameSideData ***sd, |
|
int *nb_sd, |
|
enum AVFrameSideDataType type, |
|
AVBufferRef *buf) |
|
{ |
|
if (!buf) |
|
return NULL; |
|
|
|
return add_side_data_from_buf_ext(sd, nb_sd, type, buf, buf->data, buf->size); |
|
} |
|
|
|
AVFrameSideData *av_frame_new_side_data_from_buf(AVFrame *frame, |
|
enum AVFrameSideDataType type, |
|
AVBufferRef *buf) |
|
{ |
|
return |
|
add_side_data_from_buf( |
|
&frame->side_data, &frame->nb_side_data, type, buf); |
|
} |
|
|
|
AVFrameSideData *av_frame_new_side_data(AVFrame *frame, |
|
enum AVFrameSideDataType type, |
|
size_t size) |
|
{ |
|
AVFrameSideData *ret; |
|
AVBufferRef *buf = av_buffer_alloc(size); |
|
ret = av_frame_new_side_data_from_buf(frame, type, buf); |
|
if (!ret) |
|
av_buffer_unref(&buf); |
|
return ret; |
|
} |
|
|
|
static AVFrameSideData *replace_side_data_from_buf(AVFrameSideData *dst, |
|
AVBufferRef *buf, int flags) |
|
{ |
|
if (!(flags & AV_FRAME_SIDE_DATA_FLAG_REPLACE)) |
|
return NULL; |
|
|
|
av_dict_free(&dst->metadata); |
|
av_buffer_unref(&dst->buf); |
|
dst->buf = buf; |
|
dst->data = buf->data; |
|
dst->size = buf->size; |
|
return dst; |
|
} |
|
|
|
AVFrameSideData *av_frame_side_data_new(AVFrameSideData ***sd, int *nb_sd, |
|
enum AVFrameSideDataType type, |
|
size_t size, unsigned int flags) |
|
{ |
|
const AVSideDataDescriptor *desc = av_frame_side_data_desc(type); |
|
AVBufferRef *buf = av_buffer_alloc(size); |
|
AVFrameSideData *ret = NULL; |
|
|
|
if (flags & AV_FRAME_SIDE_DATA_FLAG_UNIQUE) |
|
remove_side_data(sd, nb_sd, type); |
|
if ((!desc || !(desc->props & AV_SIDE_DATA_PROP_MULTI)) && |
|
(ret = (AVFrameSideData *)av_frame_side_data_get(*sd, *nb_sd, type))) { |
|
ret = replace_side_data_from_buf(ret, buf, flags); |
|
if (!ret) |
|
av_buffer_unref(&buf); |
|
return ret; |
|
} |
|
|
|
ret = add_side_data_from_buf(sd, nb_sd, type, buf); |
|
if (!ret) |
|
av_buffer_unref(&buf); |
|
|
|
return ret; |
|
} |
|
|
|
AVFrameSideData *av_frame_side_data_add(AVFrameSideData ***sd, int *nb_sd, |
|
enum AVFrameSideDataType type, |
|
AVBufferRef **pbuf, unsigned int flags) |
|
{ |
|
const AVSideDataDescriptor *desc = av_frame_side_data_desc(type); |
|
AVFrameSideData *sd_dst = NULL; |
|
AVBufferRef *buf = *pbuf; |
|
|
|
if (flags & AV_FRAME_SIDE_DATA_FLAG_UNIQUE) |
|
remove_side_data(sd, nb_sd, type); |
|
if ((!desc || !(desc->props & AV_SIDE_DATA_PROP_MULTI)) && |
|
(sd_dst = (AVFrameSideData *)av_frame_side_data_get(*sd, *nb_sd, type))) { |
|
sd_dst = replace_side_data_from_buf(sd_dst, buf, flags); |
|
if (sd_dst) |
|
*pbuf = NULL; |
|
return sd_dst; |
|
} |
|
|
|
sd_dst = add_side_data_from_buf(sd, nb_sd, type, buf); |
|
if (!sd_dst) |
|
return NULL; |
|
|
|
*pbuf = NULL; |
|
return sd_dst; |
|
} |
|
|
|
int av_frame_side_data_clone(AVFrameSideData ***sd, int *nb_sd, |
|
const AVFrameSideData *src, unsigned int flags) |
|
{ |
|
const AVSideDataDescriptor *desc; |
|
AVBufferRef *buf = NULL; |
|
AVFrameSideData *sd_dst = NULL; |
|
int ret = AVERROR_BUG; |
|
|
|
if (!sd || !src || !nb_sd || (*nb_sd && !*sd)) |
|
return AVERROR(EINVAL); |
|
|
|
desc = av_frame_side_data_desc(src->type); |
|
if (flags & AV_FRAME_SIDE_DATA_FLAG_UNIQUE) |
|
remove_side_data(sd, nb_sd, src->type); |
|
if ((!desc || !(desc->props & AV_SIDE_DATA_PROP_MULTI)) && |
|
(sd_dst = (AVFrameSideData *)av_frame_side_data_get(*sd, *nb_sd, src->type))) { |
|
AVDictionary *dict = NULL; |
|
|
|
if (!(flags & AV_FRAME_SIDE_DATA_FLAG_REPLACE)) |
|
return AVERROR(EEXIST); |
|
|
|
ret = av_dict_copy(&dict, src->metadata, 0); |
|
if (ret < 0) |
|
return ret; |
|
|
|
ret = av_buffer_replace(&sd_dst->buf, src->buf); |
|
if (ret < 0) { |
|
av_dict_free(&dict); |
|
return ret; |
|
} |
|
|
|
av_dict_free(&sd_dst->metadata); |
|
sd_dst->metadata = dict; |
|
sd_dst->data = src->data; |
|
sd_dst->size = src->size; |
|
return 0; |
|
} |
|
|
|
buf = av_buffer_ref(src->buf); |
|
if (!buf) |
|
return AVERROR(ENOMEM); |
|
|
|
sd_dst = add_side_data_from_buf_ext(sd, nb_sd, src->type, buf, |
|
src->data, src->size); |
|
if (!sd_dst) { |
|
av_buffer_unref(&buf); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
ret = av_dict_copy(&sd_dst->metadata, src->metadata, 0); |
|
if (ret < 0) { |
|
remove_side_data_by_entry(sd, nb_sd, sd_dst); |
|
return ret; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
const AVFrameSideData *av_frame_side_data_get_c(const AVFrameSideData * const *sd, |
|
const int nb_sd, |
|
enum AVFrameSideDataType type) |
|
{ |
|
for (int i = 0; i < nb_sd; i++) { |
|
if (sd[i]->type == type) |
|
return sd[i]; |
|
} |
|
return NULL; |
|
} |
|
|
|
void av_frame_side_data_remove(AVFrameSideData ***sd, int *nb_sd, |
|
enum AVFrameSideDataType type) |
|
{ |
|
remove_side_data(sd, nb_sd, type); |
|
} |
|
|
|
AVFrameSideData *av_frame_get_side_data(const AVFrame *frame, |
|
enum AVFrameSideDataType type) |
|
{ |
|
return (AVFrameSideData *)av_frame_side_data_get( |
|
frame->side_data, frame->nb_side_data, |
|
type |
|
); |
|
} |
|
|
|
static int frame_copy_video(AVFrame *dst, const AVFrame *src) |
|
{ |
|
int planes; |
|
|
|
if (dst->width < src->width || |
|
dst->height < src->height) |
|
return AVERROR(EINVAL); |
|
|
|
if (src->hw_frames_ctx || dst->hw_frames_ctx) |
|
return av_hwframe_transfer_data(dst, src, 0); |
|
|
|
planes = av_pix_fmt_count_planes(dst->format); |
|
for (int i = 0; i < planes; i++) |
|
if (!dst->data[i] || !src->data[i]) |
|
return AVERROR(EINVAL); |
|
|
|
av_image_copy2(dst->data, dst->linesize, |
|
src->data, src->linesize, |
|
dst->format, src->width, src->height); |
|
|
|
return 0; |
|
} |
|
|
|
static int frame_copy_audio(AVFrame *dst, const AVFrame *src) |
|
{ |
|
int planar = av_sample_fmt_is_planar(dst->format); |
|
int channels = dst->ch_layout.nb_channels; |
|
int planes = planar ? channels : 1; |
|
|
|
if (dst->nb_samples != src->nb_samples || |
|
av_channel_layout_compare(&dst->ch_layout, &src->ch_layout)) |
|
return AVERROR(EINVAL); |
|
|
|
for (int i = 0; i < planes; i++) |
|
if (!dst->extended_data[i] || !src->extended_data[i]) |
|
return AVERROR(EINVAL); |
|
|
|
av_samples_copy(dst->extended_data, src->extended_data, 0, 0, |
|
dst->nb_samples, channels, dst->format); |
|
|
|
return 0; |
|
} |
|
|
|
int av_frame_copy(AVFrame *dst, const AVFrame *src) |
|
{ |
|
if (dst->format != src->format || dst->format < 0) |
|
return AVERROR(EINVAL); |
|
|
|
if (dst->width > 0 && dst->height > 0) |
|
return frame_copy_video(dst, src); |
|
else if (dst->nb_samples > 0 && |
|
(av_channel_layout_check(&dst->ch_layout))) |
|
return frame_copy_audio(dst, src); |
|
|
|
return AVERROR(EINVAL); |
|
} |
|
|
|
void av_frame_remove_side_data(AVFrame *frame, enum AVFrameSideDataType type) |
|
{ |
|
remove_side_data(&frame->side_data, &frame->nb_side_data, type); |
|
} |
|
|
|
const AVSideDataDescriptor *av_frame_side_data_desc(enum AVFrameSideDataType type) |
|
{ |
|
unsigned t = type; |
|
if (t < FF_ARRAY_ELEMS(sd_props) && sd_props[t].name) |
|
return &sd_props[t]; |
|
return NULL; |
|
} |
|
|
|
const char *av_frame_side_data_name(enum AVFrameSideDataType type) |
|
{ |
|
const AVSideDataDescriptor *desc = av_frame_side_data_desc(type); |
|
return desc ? desc->name : NULL; |
|
} |
|
|
|
static int calc_cropping_offsets(size_t offsets[4], const AVFrame *frame, |
|
const AVPixFmtDescriptor *desc) |
|
{ |
|
for (int i = 0; frame->data[i]; i++) { |
|
const AVComponentDescriptor *comp = NULL; |
|
int shift_x = (i == 1 || i == 2) ? desc->log2_chroma_w : 0; |
|
int shift_y = (i == 1 || i == 2) ? desc->log2_chroma_h : 0; |
|
|
|
if (desc->flags & AV_PIX_FMT_FLAG_PAL && i == 1) { |
|
offsets[i] = 0; |
|
break; |
|
} |
|
|
|
/* find any component descriptor for this plane */ |
|
for (int j = 0; j < desc->nb_components; j++) { |
|
if (desc->comp[j].plane == i) { |
|
comp = &desc->comp[j]; |
|
break; |
|
} |
|
} |
|
if (!comp) |
|
return AVERROR_BUG; |
|
|
|
offsets[i] = (frame->crop_top >> shift_y) * frame->linesize[i] + |
|
(frame->crop_left >> shift_x) * comp->step; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
int av_frame_apply_cropping(AVFrame *frame, int flags) |
|
{ |
|
const AVPixFmtDescriptor *desc; |
|
size_t offsets[4]; |
|
|
|
if (!(frame->width > 0 && frame->height > 0)) |
|
return AVERROR(EINVAL); |
|
|
|
if (frame->crop_left >= INT_MAX - frame->crop_right || |
|
frame->crop_top >= INT_MAX - frame->crop_bottom || |
|
(frame->crop_left + frame->crop_right) >= frame->width || |
|
(frame->crop_top + frame->crop_bottom) >= frame->height) |
|
return AVERROR(ERANGE); |
|
|
|
desc = av_pix_fmt_desc_get(frame->format); |
|
if (!desc) |
|
return AVERROR_BUG; |
|
|
|
/* Apply just the right/bottom cropping for hwaccel formats. Bitstream |
|
* formats cannot be easily handled here either (and corresponding decoders |
|
* should not export any cropping anyway), so do the same for those as well. |
|
* */ |
|
if (desc->flags & (AV_PIX_FMT_FLAG_BITSTREAM | AV_PIX_FMT_FLAG_HWACCEL)) { |
|
frame->width -= frame->crop_right; |
|
frame->height -= frame->crop_bottom; |
|
frame->crop_right = 0; |
|
frame->crop_bottom = 0; |
|
return 0; |
|
} |
|
|
|
/* calculate the offsets for each plane */ |
|
calc_cropping_offsets(offsets, frame, desc); |
|
|
|
/* adjust the offsets to avoid breaking alignment */ |
|
if (!(flags & AV_FRAME_CROP_UNALIGNED)) { |
|
int log2_crop_align = frame->crop_left ? ff_ctz(frame->crop_left) : INT_MAX; |
|
int min_log2_align = INT_MAX; |
|
|
|
for (int i = 0; frame->data[i]; i++) { |
|
int log2_align = offsets[i] ? ff_ctz(offsets[i]) : INT_MAX; |
|
min_log2_align = FFMIN(log2_align, min_log2_align); |
|
} |
|
|
|
/* we assume, and it should always be true, that the data alignment is |
|
* related to the cropping alignment by a constant power-of-2 factor */ |
|
if (log2_crop_align < min_log2_align) |
|
return AVERROR_BUG; |
|
|
|
if (min_log2_align < 5 && log2_crop_align != INT_MAX) { |
|
frame->crop_left &= ~((1 << (5 + log2_crop_align - min_log2_align)) - 1); |
|
calc_cropping_offsets(offsets, frame, desc); |
|
} |
|
} |
|
|
|
for (int i = 0; frame->data[i]; i++) |
|
frame->data[i] += offsets[i]; |
|
|
|
frame->width -= (frame->crop_left + frame->crop_right); |
|
frame->height -= (frame->crop_top + frame->crop_bottom); |
|
frame->crop_left = 0; |
|
frame->crop_right = 0; |
|
frame->crop_top = 0; |
|
frame->crop_bottom = 0; |
|
|
|
return 0; |
|
}
|
|
|