|
|
@ -166,6 +166,13 @@ struct TrellisNode { |
|
|
|
|
|
|
|
|
|
|
|
#define MAX_STORED_Q 16 |
|
|
|
#define MAX_STORED_Q 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
typedef struct ProresThreadData { |
|
|
|
|
|
|
|
DECLARE_ALIGNED(16, DCTELEM, blocks)[MAX_PLANES][64 * 4 * MAX_MBS_PER_SLICE]; |
|
|
|
|
|
|
|
DECLARE_ALIGNED(16, uint16_t, emu_buf)[16 * 16]; |
|
|
|
|
|
|
|
int16_t custom_q[64]; |
|
|
|
|
|
|
|
struct TrellisNode *nodes; |
|
|
|
|
|
|
|
} ProresThreadData; |
|
|
|
|
|
|
|
|
|
|
|
typedef struct ProresContext { |
|
|
|
typedef struct ProresContext { |
|
|
|
AVClass *class; |
|
|
|
AVClass *class; |
|
|
|
DECLARE_ALIGNED(16, DCTELEM, blocks)[MAX_PLANES][64 * 4 * MAX_MBS_PER_SLICE]; |
|
|
|
DECLARE_ALIGNED(16, DCTELEM, blocks)[MAX_PLANES][64 * 4 * MAX_MBS_PER_SLICE]; |
|
|
@ -194,13 +201,14 @@ typedef struct ProresContext { |
|
|
|
int profile; |
|
|
|
int profile; |
|
|
|
const struct prores_profile *profile_info; |
|
|
|
const struct prores_profile *profile_info; |
|
|
|
|
|
|
|
|
|
|
|
struct TrellisNode *nodes; |
|
|
|
|
|
|
|
int *slice_q; |
|
|
|
int *slice_q; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
ProresThreadData *tdata; |
|
|
|
} ProresContext; |
|
|
|
} ProresContext; |
|
|
|
|
|
|
|
|
|
|
|
static void get_slice_data(ProresContext *ctx, const uint16_t *src, |
|
|
|
static void get_slice_data(ProresContext *ctx, const uint16_t *src, |
|
|
|
int linesize, int x, int y, int w, int h, |
|
|
|
int linesize, int x, int y, int w, int h, |
|
|
|
DCTELEM *blocks, |
|
|
|
DCTELEM *blocks, uint16_t *emu_buf, |
|
|
|
int mbs_per_slice, int blocks_per_mb, int is_chroma) |
|
|
|
int mbs_per_slice, int blocks_per_mb, int is_chroma) |
|
|
|
{ |
|
|
|
{ |
|
|
|
const uint16_t *esrc; |
|
|
|
const uint16_t *esrc; |
|
|
@ -220,24 +228,24 @@ static void get_slice_data(ProresContext *ctx, const uint16_t *src, |
|
|
|
} else { |
|
|
|
} else { |
|
|
|
int bw, bh, pix; |
|
|
|
int bw, bh, pix; |
|
|
|
|
|
|
|
|
|
|
|
esrc = ctx->emu_buf; |
|
|
|
esrc = emu_buf; |
|
|
|
elinesize = 16 * sizeof(*ctx->emu_buf); |
|
|
|
elinesize = 16 * sizeof(*emu_buf); |
|
|
|
|
|
|
|
|
|
|
|
bw = FFMIN(w - x, mb_width); |
|
|
|
bw = FFMIN(w - x, mb_width); |
|
|
|
bh = FFMIN(h - y, 16); |
|
|
|
bh = FFMIN(h - y, 16); |
|
|
|
|
|
|
|
|
|
|
|
for (j = 0; j < bh; j++) { |
|
|
|
for (j = 0; j < bh; j++) { |
|
|
|
memcpy(ctx->emu_buf + j * 16, |
|
|
|
memcpy(emu_buf + j * 16, |
|
|
|
(const uint8_t*)src + j * linesize, |
|
|
|
(const uint8_t*)src + j * linesize, |
|
|
|
bw * sizeof(*src)); |
|
|
|
bw * sizeof(*src)); |
|
|
|
pix = ctx->emu_buf[j * 16 + bw - 1]; |
|
|
|
pix = emu_buf[j * 16 + bw - 1]; |
|
|
|
for (k = bw; k < mb_width; k++) |
|
|
|
for (k = bw; k < mb_width; k++) |
|
|
|
ctx->emu_buf[j * 16 + k] = pix; |
|
|
|
emu_buf[j * 16 + k] = pix; |
|
|
|
} |
|
|
|
} |
|
|
|
for (; j < 16; j++) |
|
|
|
for (; j < 16; j++) |
|
|
|
memcpy(ctx->emu_buf + j * 16, |
|
|
|
memcpy(emu_buf + j * 16, |
|
|
|
ctx->emu_buf + (bh - 1) * 16, |
|
|
|
emu_buf + (bh - 1) * 16, |
|
|
|
mb_width * sizeof(*ctx->emu_buf)); |
|
|
|
mb_width * sizeof(*emu_buf)); |
|
|
|
} |
|
|
|
} |
|
|
|
if (!is_chroma) { |
|
|
|
if (!is_chroma) { |
|
|
|
ctx->dsp.fdct(esrc, elinesize, blocks); |
|
|
|
ctx->dsp.fdct(esrc, elinesize, blocks); |
|
|
@ -427,7 +435,7 @@ static int encode_slice(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
src = (const uint16_t*)(pic->data[i] + yp * pic->linesize[i]) + xp; |
|
|
|
src = (const uint16_t*)(pic->data[i] + yp * pic->linesize[i]) + xp; |
|
|
|
|
|
|
|
|
|
|
|
get_slice_data(ctx, src, pic->linesize[i], xp, yp, |
|
|
|
get_slice_data(ctx, src, pic->linesize[i], xp, yp, |
|
|
|
pwidth, avctx->height, ctx->blocks[0], |
|
|
|
pwidth, avctx->height, ctx->blocks[0], ctx->emu_buf, |
|
|
|
mbs_per_slice, num_cblocks, is_chroma); |
|
|
|
mbs_per_slice, num_cblocks, is_chroma); |
|
|
|
sizes[i] = encode_slice_plane(ctx, pb, src, pic->linesize[i], |
|
|
|
sizes[i] = encode_slice_plane(ctx, pb, src, pic->linesize[i], |
|
|
|
mbs_per_slice, ctx->blocks[0], |
|
|
|
mbs_per_slice, ctx->blocks[0], |
|
|
@ -531,22 +539,23 @@ static int estimate_slice_plane(ProresContext *ctx, int *error, int plane, |
|
|
|
const uint16_t *src, int linesize, |
|
|
|
const uint16_t *src, int linesize, |
|
|
|
int mbs_per_slice, |
|
|
|
int mbs_per_slice, |
|
|
|
int blocks_per_mb, int plane_size_factor, |
|
|
|
int blocks_per_mb, int plane_size_factor, |
|
|
|
const int16_t *qmat) |
|
|
|
const int16_t *qmat, ProresThreadData *td) |
|
|
|
{ |
|
|
|
{ |
|
|
|
int blocks_per_slice; |
|
|
|
int blocks_per_slice; |
|
|
|
int bits; |
|
|
|
int bits; |
|
|
|
|
|
|
|
|
|
|
|
blocks_per_slice = mbs_per_slice * blocks_per_mb; |
|
|
|
blocks_per_slice = mbs_per_slice * blocks_per_mb; |
|
|
|
|
|
|
|
|
|
|
|
bits = estimate_dcs(error, ctx->blocks[plane], blocks_per_slice, qmat[0]); |
|
|
|
bits = estimate_dcs(error, td->blocks[plane], blocks_per_slice, qmat[0]); |
|
|
|
bits += estimate_acs(error, ctx->blocks[plane], blocks_per_slice, |
|
|
|
bits += estimate_acs(error, td->blocks[plane], blocks_per_slice, |
|
|
|
plane_size_factor, ctx->scantable.permutated, qmat); |
|
|
|
plane_size_factor, ctx->scantable.permutated, qmat); |
|
|
|
|
|
|
|
|
|
|
|
return FFALIGN(bits, 8); |
|
|
|
return FFALIGN(bits, 8); |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
int trellis_node, int x, int y, int mbs_per_slice) |
|
|
|
int trellis_node, int x, int y, int mbs_per_slice, |
|
|
|
|
|
|
|
ProresThreadData *td) |
|
|
|
{ |
|
|
|
{ |
|
|
|
ProresContext *ctx = avctx->priv_data; |
|
|
|
ProresContext *ctx = avctx->priv_data; |
|
|
|
int i, q, pq, xp, yp; |
|
|
|
int i, q, pq, xp, yp; |
|
|
@ -583,13 +592,13 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
src = (const uint16_t*)(pic->data[i] + yp * pic->linesize[i]) + xp; |
|
|
|
src = (const uint16_t*)(pic->data[i] + yp * pic->linesize[i]) + xp; |
|
|
|
|
|
|
|
|
|
|
|
get_slice_data(ctx, src, pic->linesize[i], xp, yp, |
|
|
|
get_slice_data(ctx, src, pic->linesize[i], xp, yp, |
|
|
|
pwidth, avctx->height, ctx->blocks[i], |
|
|
|
pwidth, avctx->height, td->blocks[i], td->emu_buf, |
|
|
|
mbs_per_slice, num_cblocks[i], is_chroma[i]); |
|
|
|
mbs_per_slice, num_cblocks[i], is_chroma[i]); |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
for (q = min_quant; q < max_quant + 2; q++) { |
|
|
|
for (q = min_quant; q < max_quant + 2; q++) { |
|
|
|
ctx->nodes[trellis_node + q].prev_node = -1; |
|
|
|
td->nodes[trellis_node + q].prev_node = -1; |
|
|
|
ctx->nodes[trellis_node + q].quant = q; |
|
|
|
td->nodes[trellis_node + q].quant = q; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
// todo: maybe perform coarser quantising to fit into frame size when needed
|
|
|
|
// todo: maybe perform coarser quantising to fit into frame size when needed
|
|
|
@ -601,7 +610,7 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
src, pic->linesize[i], |
|
|
|
src, pic->linesize[i], |
|
|
|
mbs_per_slice, |
|
|
|
mbs_per_slice, |
|
|
|
num_cblocks[i], plane_factor[i], |
|
|
|
num_cblocks[i], plane_factor[i], |
|
|
|
ctx->quants[q]); |
|
|
|
ctx->quants[q], td); |
|
|
|
} |
|
|
|
} |
|
|
|
if (bits > 65000 * 8) { |
|
|
|
if (bits > 65000 * 8) { |
|
|
|
error = SCORE_LIMIT; |
|
|
|
error = SCORE_LIMIT; |
|
|
@ -621,7 +630,7 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
if (q < MAX_STORED_Q) { |
|
|
|
if (q < MAX_STORED_Q) { |
|
|
|
qmat = ctx->quants[q]; |
|
|
|
qmat = ctx->quants[q]; |
|
|
|
} else { |
|
|
|
} else { |
|
|
|
qmat = ctx->custom_q; |
|
|
|
qmat = td->custom_q; |
|
|
|
for (i = 0; i < 64; i++) |
|
|
|
for (i = 0; i < 64; i++) |
|
|
|
qmat[i] = ctx->quant_mat[i] * q; |
|
|
|
qmat[i] = ctx->quant_mat[i] * q; |
|
|
|
} |
|
|
|
} |
|
|
@ -630,7 +639,7 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
src, pic->linesize[i], |
|
|
|
src, pic->linesize[i], |
|
|
|
mbs_per_slice, |
|
|
|
mbs_per_slice, |
|
|
|
num_cblocks[i], plane_factor[i], |
|
|
|
num_cblocks[i], plane_factor[i], |
|
|
|
qmat); |
|
|
|
qmat, td); |
|
|
|
} |
|
|
|
} |
|
|
|
if (bits <= ctx->bits_per_mb * mbs_per_slice) |
|
|
|
if (bits <= ctx->bits_per_mb * mbs_per_slice) |
|
|
|
break; |
|
|
|
break; |
|
|
@ -640,7 +649,7 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
slice_score[max_quant + 1] = error; |
|
|
|
slice_score[max_quant + 1] = error; |
|
|
|
overquant = q; |
|
|
|
overquant = q; |
|
|
|
} |
|
|
|
} |
|
|
|
ctx->nodes[trellis_node + max_quant + 1].quant = overquant; |
|
|
|
td->nodes[trellis_node + max_quant + 1].quant = overquant; |
|
|
|
|
|
|
|
|
|
|
|
bits_limit = mbs * ctx->bits_per_mb; |
|
|
|
bits_limit = mbs * ctx->bits_per_mb; |
|
|
|
for (pq = min_quant; pq < max_quant + 2; pq++) { |
|
|
|
for (pq = min_quant; pq < max_quant + 2; pq++) { |
|
|
@ -649,30 +658,30 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
for (q = min_quant; q < max_quant + 2; q++) { |
|
|
|
for (q = min_quant; q < max_quant + 2; q++) { |
|
|
|
cur = trellis_node + q; |
|
|
|
cur = trellis_node + q; |
|
|
|
|
|
|
|
|
|
|
|
bits = ctx->nodes[prev].bits + slice_bits[q]; |
|
|
|
bits = td->nodes[prev].bits + slice_bits[q]; |
|
|
|
error = slice_score[q]; |
|
|
|
error = slice_score[q]; |
|
|
|
if (bits > bits_limit) |
|
|
|
if (bits > bits_limit) |
|
|
|
error = SCORE_LIMIT; |
|
|
|
error = SCORE_LIMIT; |
|
|
|
|
|
|
|
|
|
|
|
if (ctx->nodes[prev].score < SCORE_LIMIT && error < SCORE_LIMIT) |
|
|
|
if (td->nodes[prev].score < SCORE_LIMIT && error < SCORE_LIMIT) |
|
|
|
new_score = ctx->nodes[prev].score + error; |
|
|
|
new_score = td->nodes[prev].score + error; |
|
|
|
else |
|
|
|
else |
|
|
|
new_score = SCORE_LIMIT; |
|
|
|
new_score = SCORE_LIMIT; |
|
|
|
if (ctx->nodes[cur].prev_node == -1 || |
|
|
|
if (td->nodes[cur].prev_node == -1 || |
|
|
|
ctx->nodes[cur].score >= new_score) { |
|
|
|
td->nodes[cur].score >= new_score) { |
|
|
|
|
|
|
|
|
|
|
|
ctx->nodes[cur].bits = bits; |
|
|
|
td->nodes[cur].bits = bits; |
|
|
|
ctx->nodes[cur].score = new_score; |
|
|
|
td->nodes[cur].score = new_score; |
|
|
|
ctx->nodes[cur].prev_node = prev; |
|
|
|
td->nodes[cur].prev_node = prev; |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
error = ctx->nodes[trellis_node + min_quant].score; |
|
|
|
error = td->nodes[trellis_node + min_quant].score; |
|
|
|
pq = trellis_node + min_quant; |
|
|
|
pq = trellis_node + min_quant; |
|
|
|
for (q = min_quant + 1; q < max_quant + 2; q++) { |
|
|
|
for (q = min_quant + 1; q < max_quant + 2; q++) { |
|
|
|
if (ctx->nodes[trellis_node + q].score <= error) { |
|
|
|
if (td->nodes[trellis_node + q].score <= error) { |
|
|
|
error = ctx->nodes[trellis_node + q].score; |
|
|
|
error = td->nodes[trellis_node + q].score; |
|
|
|
pq = trellis_node + q; |
|
|
|
pq = trellis_node + q; |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
|
} |
|
|
@ -680,6 +689,30 @@ static int find_slice_quant(AVCodecContext *avctx, const AVFrame *pic, |
|
|
|
return pq; |
|
|
|
return pq; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int find_quant_thread(AVCodecContext *avctx, void *arg, |
|
|
|
|
|
|
|
int jobnr, int threadnr) |
|
|
|
|
|
|
|
{ |
|
|
|
|
|
|
|
ProresContext *ctx = avctx->priv_data; |
|
|
|
|
|
|
|
ProresThreadData *td = ctx->tdata + threadnr; |
|
|
|
|
|
|
|
int mbs_per_slice = ctx->mbs_per_slice; |
|
|
|
|
|
|
|
int x, y = jobnr, mb, q = 0; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (x = mb = 0; x < ctx->mb_width; x += mbs_per_slice, mb++) { |
|
|
|
|
|
|
|
while (ctx->mb_width - x < mbs_per_slice) |
|
|
|
|
|
|
|
mbs_per_slice >>= 1; |
|
|
|
|
|
|
|
q = find_slice_quant(avctx, avctx->coded_frame, |
|
|
|
|
|
|
|
(mb + 1) * TRELLIS_WIDTH, x, y, |
|
|
|
|
|
|
|
mbs_per_slice, td); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (x = ctx->slices_width - 1; x >= 0; x--) { |
|
|
|
|
|
|
|
ctx->slice_q[x + y * ctx->slices_width] = td->nodes[q].quant; |
|
|
|
|
|
|
|
q = td->nodes[q].prev_node; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return 0; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, |
|
|
|
static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, |
|
|
|
const AVFrame *pic, int *got_packet) |
|
|
|
const AVFrame *pic, int *got_packet) |
|
|
|
{ |
|
|
|
{ |
|
|
@ -751,25 +784,18 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, |
|
|
|
buf += ctx->num_slices * 2; |
|
|
|
buf += ctx->num_slices * 2; |
|
|
|
|
|
|
|
|
|
|
|
// slices
|
|
|
|
// slices
|
|
|
|
for (y = 0; y < ctx->mb_height; y++) { |
|
|
|
|
|
|
|
mbs_per_slice = ctx->mbs_per_slice; |
|
|
|
|
|
|
|
if (!ctx->force_quant) { |
|
|
|
if (!ctx->force_quant) { |
|
|
|
for (x = mb = 0; x < ctx->mb_width; x += mbs_per_slice, mb++) { |
|
|
|
ret = avctx->execute2(avctx, find_quant_thread, NULL, NULL, |
|
|
|
while (ctx->mb_width - x < mbs_per_slice) |
|
|
|
ctx->mb_height); |
|
|
|
mbs_per_slice >>= 1; |
|
|
|
if (ret) |
|
|
|
q = find_slice_quant(avctx, pic, (mb + 1) * TRELLIS_WIDTH, x, y, |
|
|
|
return ret; |
|
|
|
mbs_per_slice); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (x = ctx->slices_width - 1; x >= 0; x--) { |
|
|
|
|
|
|
|
ctx->slice_q[x] = ctx->nodes[q].quant; |
|
|
|
|
|
|
|
q = ctx->nodes[q].prev_node; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (y = 0; y < ctx->mb_height; y++) { |
|
|
|
mbs_per_slice = ctx->mbs_per_slice; |
|
|
|
mbs_per_slice = ctx->mbs_per_slice; |
|
|
|
for (x = mb = 0; x < ctx->mb_width; x += mbs_per_slice, mb++) { |
|
|
|
for (x = mb = 0; x < ctx->mb_width; x += mbs_per_slice, mb++) { |
|
|
|
q = ctx->force_quant ? ctx->force_quant : ctx->slice_q[mb]; |
|
|
|
q = ctx->force_quant ? ctx->force_quant |
|
|
|
|
|
|
|
: ctx->slice_q[mb + y * ctx->slices_width]; |
|
|
|
|
|
|
|
|
|
|
|
while (ctx->mb_width - x < mbs_per_slice) |
|
|
|
while (ctx->mb_width - x < mbs_per_slice) |
|
|
|
mbs_per_slice >>= 1; |
|
|
|
mbs_per_slice >>= 1; |
|
|
@ -807,13 +833,18 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, |
|
|
|
static av_cold int encode_close(AVCodecContext *avctx) |
|
|
|
static av_cold int encode_close(AVCodecContext *avctx) |
|
|
|
{ |
|
|
|
{ |
|
|
|
ProresContext *ctx = avctx->priv_data; |
|
|
|
ProresContext *ctx = avctx->priv_data; |
|
|
|
|
|
|
|
int i; |
|
|
|
|
|
|
|
|
|
|
|
if (avctx->coded_frame->data[0]) |
|
|
|
if (avctx->coded_frame->data[0]) |
|
|
|
avctx->release_buffer(avctx, avctx->coded_frame); |
|
|
|
avctx->release_buffer(avctx, avctx->coded_frame); |
|
|
|
|
|
|
|
|
|
|
|
av_freep(&avctx->coded_frame); |
|
|
|
av_freep(&avctx->coded_frame); |
|
|
|
|
|
|
|
|
|
|
|
av_freep(&ctx->nodes); |
|
|
|
if (ctx->tdata) { |
|
|
|
|
|
|
|
for (i = 0; i < avctx->thread_count; i++) |
|
|
|
|
|
|
|
av_free(ctx->tdata[i].nodes); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
av_freep(&ctx->tdata); |
|
|
|
av_freep(&ctx->slice_q); |
|
|
|
av_freep(&ctx->slice_q); |
|
|
|
|
|
|
|
|
|
|
|
return 0; |
|
|
|
return 0; |
|
|
@ -883,23 +914,32 @@ static av_cold int encode_init(AVCodecContext *avctx) |
|
|
|
ctx->quants[i][j] = ctx->quant_mat[j] * i; |
|
|
|
ctx->quants[i][j] = ctx->quant_mat[j] * i; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
ctx->nodes = av_malloc((ctx->slices_width + 1) * TRELLIS_WIDTH |
|
|
|
ctx->slice_q = av_malloc(ctx->num_slices * sizeof(*ctx->slice_q)); |
|
|
|
* sizeof(*ctx->nodes)); |
|
|
|
if (!ctx->slice_q) { |
|
|
|
if (!ctx->nodes) { |
|
|
|
|
|
|
|
encode_close(avctx); |
|
|
|
encode_close(avctx); |
|
|
|
return AVERROR(ENOMEM); |
|
|
|
return AVERROR(ENOMEM); |
|
|
|
} |
|
|
|
} |
|
|
|
for (i = min_quant; i < max_quant + 2; i++) { |
|
|
|
|
|
|
|
ctx->nodes[i].prev_node = -1; |
|
|
|
ctx->tdata = av_mallocz(avctx->thread_count * sizeof(*ctx->tdata)); |
|
|
|
ctx->nodes[i].bits = 0; |
|
|
|
if (!ctx->tdata) { |
|
|
|
ctx->nodes[i].score = 0; |
|
|
|
encode_close(avctx); |
|
|
|
|
|
|
|
return AVERROR(ENOMEM); |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
ctx->slice_q = av_malloc(ctx->slices_width * sizeof(*ctx->slice_q)); |
|
|
|
for (j = 0; j < avctx->thread_count; j++) { |
|
|
|
if (!ctx->slice_q) { |
|
|
|
ctx->tdata[j].nodes = av_malloc((ctx->slices_width + 1) |
|
|
|
|
|
|
|
* TRELLIS_WIDTH |
|
|
|
|
|
|
|
* sizeof(*ctx->tdata->nodes)); |
|
|
|
|
|
|
|
if (!ctx->tdata[j].nodes) { |
|
|
|
encode_close(avctx); |
|
|
|
encode_close(avctx); |
|
|
|
return AVERROR(ENOMEM); |
|
|
|
return AVERROR(ENOMEM); |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
for (i = min_quant; i < max_quant + 2; i++) { |
|
|
|
|
|
|
|
ctx->tdata[j].nodes[i].prev_node = -1; |
|
|
|
|
|
|
|
ctx->tdata[j].nodes[i].bits = 0; |
|
|
|
|
|
|
|
ctx->tdata[j].nodes[i].score = 0; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
} |
|
|
|
} else { |
|
|
|
} else { |
|
|
|
int ls = 0; |
|
|
|
int ls = 0; |
|
|
|
|
|
|
|
|
|
|
@ -987,6 +1027,7 @@ AVCodec ff_prores_encoder = { |
|
|
|
.init = encode_init, |
|
|
|
.init = encode_init, |
|
|
|
.close = encode_close, |
|
|
|
.close = encode_close, |
|
|
|
.encode2 = encode_frame, |
|
|
|
.encode2 = encode_frame, |
|
|
|
|
|
|
|
.capabilities = CODEC_CAP_SLICE_THREADS, |
|
|
|
.long_name = NULL_IF_CONFIG_SMALL("Apple ProRes (iCodec Pro)"), |
|
|
|
.long_name = NULL_IF_CONFIG_SMALL("Apple ProRes (iCodec Pro)"), |
|
|
|
.pix_fmts = (const enum PixelFormat[]) { |
|
|
|
.pix_fmts = (const enum PixelFormat[]) { |
|
|
|
PIX_FMT_YUV422P10, PIX_FMT_YUV444P10, PIX_FMT_NONE |
|
|
|
PIX_FMT_YUV422P10, PIX_FMT_YUV444P10, PIX_FMT_NONE |
|
|
|