mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
673 lines
24 KiB
673 lines
24 KiB
/* |
|
* JPEG 2000 encoding support via OpenJPEG |
|
* Copyright (c) 2011 Michael Bradshaw <mjbshaw gmail com> |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
/** |
|
* @file |
|
* JPEG 2000 encoder using libopenjpeg |
|
*/ |
|
|
|
#define OPJ_STATIC |
|
|
|
#include "libavutil/avassert.h" |
|
#include "libavutil/common.h" |
|
#include "libavutil/imgutils.h" |
|
#include "libavutil/intreadwrite.h" |
|
#include "libavutil/opt.h" |
|
#include "avcodec.h" |
|
#include "internal.h" |
|
|
|
#if HAVE_OPENJPEG_1_5_OPENJPEG_H |
|
# include <openjpeg-1.5/openjpeg.h> |
|
#else |
|
# include <openjpeg.h> |
|
#endif |
|
|
|
typedef struct LibOpenJPEGContext { |
|
AVClass *avclass; |
|
opj_image_t *image; |
|
opj_cparameters_t enc_params; |
|
opj_event_mgr_t event_mgr; |
|
int format; |
|
int profile; |
|
int prog_order; |
|
int cinema_mode; |
|
int numresolution; |
|
int numlayers; |
|
int disto_alloc; |
|
int fixed_alloc; |
|
int fixed_quality; |
|
} LibOpenJPEGContext; |
|
|
|
static void error_callback(const char *msg, void *data) |
|
{ |
|
av_log(data, AV_LOG_ERROR, "%s\n", msg); |
|
} |
|
|
|
static void warning_callback(const char *msg, void *data) |
|
{ |
|
av_log(data, AV_LOG_WARNING, "%s\n", msg); |
|
} |
|
|
|
static void info_callback(const char *msg, void *data) |
|
{ |
|
av_log(data, AV_LOG_DEBUG, "%s\n", msg); |
|
} |
|
|
|
static void cinema_parameters(opj_cparameters_t *p) |
|
{ |
|
p->tile_size_on = 0; |
|
p->cp_tdx = 1; |
|
p->cp_tdy = 1; |
|
|
|
/* Tile part */ |
|
p->tp_flag = 'C'; |
|
p->tp_on = 1; |
|
|
|
/* Tile and Image shall be at (0, 0) */ |
|
p->cp_tx0 = 0; |
|
p->cp_ty0 = 0; |
|
p->image_offset_x0 = 0; |
|
p->image_offset_y0 = 0; |
|
|
|
/* Codeblock size= 32 * 32 */ |
|
p->cblockw_init = 32; |
|
p->cblockh_init = 32; |
|
p->csty |= 0x01; |
|
|
|
/* The progression order shall be CPRL */ |
|
p->prog_order = CPRL; |
|
|
|
/* No ROI */ |
|
p->roi_compno = -1; |
|
|
|
/* No subsampling */ |
|
p->subsampling_dx = 1; |
|
p->subsampling_dy = 1; |
|
|
|
/* 9-7 transform */ |
|
p->irreversible = 1; |
|
|
|
p->tcp_mct = 1; |
|
} |
|
|
|
static opj_image_t *mj2_create_image(AVCodecContext *avctx, opj_cparameters_t *parameters) |
|
{ |
|
const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(avctx->pix_fmt); |
|
opj_image_cmptparm_t cmptparm[4] = {{0}}; |
|
opj_image_t *img; |
|
int i; |
|
int sub_dx[4]; |
|
int sub_dy[4]; |
|
int numcomps; |
|
OPJ_COLOR_SPACE color_space = CLRSPC_UNKNOWN; |
|
|
|
sub_dx[0] = sub_dx[3] = 1; |
|
sub_dy[0] = sub_dy[3] = 1; |
|
sub_dx[1] = sub_dx[2] = 1 << desc->log2_chroma_w; |
|
sub_dy[1] = sub_dy[2] = 1 << desc->log2_chroma_h; |
|
|
|
numcomps = desc->nb_components; |
|
|
|
switch (avctx->pix_fmt) { |
|
case AV_PIX_FMT_GRAY8: |
|
case AV_PIX_FMT_YA8: |
|
case AV_PIX_FMT_GRAY16: |
|
case AV_PIX_FMT_YA16: |
|
color_space = CLRSPC_GRAY; |
|
break; |
|
case AV_PIX_FMT_RGB24: |
|
case AV_PIX_FMT_RGBA: |
|
case AV_PIX_FMT_RGB48: |
|
case AV_PIX_FMT_RGBA64: |
|
case AV_PIX_FMT_GBR24P: |
|
case AV_PIX_FMT_GBRP9: |
|
case AV_PIX_FMT_GBRP10: |
|
case AV_PIX_FMT_GBRP12: |
|
case AV_PIX_FMT_GBRP14: |
|
case AV_PIX_FMT_GBRP16: |
|
case AV_PIX_FMT_XYZ12: |
|
color_space = CLRSPC_SRGB; |
|
break; |
|
case AV_PIX_FMT_YUV410P: |
|
case AV_PIX_FMT_YUV411P: |
|
case AV_PIX_FMT_YUV420P: |
|
case AV_PIX_FMT_YUV422P: |
|
case AV_PIX_FMT_YUV440P: |
|
case AV_PIX_FMT_YUV444P: |
|
case AV_PIX_FMT_YUVA420P: |
|
case AV_PIX_FMT_YUVA422P: |
|
case AV_PIX_FMT_YUVA444P: |
|
case AV_PIX_FMT_YUV420P9: |
|
case AV_PIX_FMT_YUV422P9: |
|
case AV_PIX_FMT_YUV444P9: |
|
case AV_PIX_FMT_YUVA420P9: |
|
case AV_PIX_FMT_YUVA422P9: |
|
case AV_PIX_FMT_YUVA444P9: |
|
case AV_PIX_FMT_YUV420P10: |
|
case AV_PIX_FMT_YUV422P10: |
|
case AV_PIX_FMT_YUV444P10: |
|
case AV_PIX_FMT_YUVA420P10: |
|
case AV_PIX_FMT_YUVA422P10: |
|
case AV_PIX_FMT_YUVA444P10: |
|
case AV_PIX_FMT_YUV420P12: |
|
case AV_PIX_FMT_YUV422P12: |
|
case AV_PIX_FMT_YUV444P12: |
|
case AV_PIX_FMT_YUV420P14: |
|
case AV_PIX_FMT_YUV422P14: |
|
case AV_PIX_FMT_YUV444P14: |
|
case AV_PIX_FMT_YUV420P16: |
|
case AV_PIX_FMT_YUV422P16: |
|
case AV_PIX_FMT_YUV444P16: |
|
case AV_PIX_FMT_YUVA420P16: |
|
case AV_PIX_FMT_YUVA422P16: |
|
case AV_PIX_FMT_YUVA444P16: |
|
color_space = CLRSPC_SYCC; |
|
break; |
|
default: |
|
av_log(avctx, AV_LOG_ERROR, |
|
"The requested pixel format '%s' is not supported\n", |
|
av_get_pix_fmt_name(avctx->pix_fmt)); |
|
return NULL; |
|
} |
|
|
|
for (i = 0; i < numcomps; i++) { |
|
cmptparm[i].prec = desc->comp[i].depth_minus1 + 1; |
|
cmptparm[i].bpp = desc->comp[i].depth_minus1 + 1; |
|
cmptparm[i].sgnd = 0; |
|
cmptparm[i].dx = sub_dx[i]; |
|
cmptparm[i].dy = sub_dy[i]; |
|
cmptparm[i].w = (avctx->width + sub_dx[i] - 1) / sub_dx[i]; |
|
cmptparm[i].h = (avctx->height + sub_dy[i] - 1) / sub_dy[i]; |
|
} |
|
|
|
img = opj_image_create(numcomps, cmptparm, color_space); |
|
|
|
if (!img) |
|
return NULL; |
|
|
|
// x0, y0 is the top left corner of the image |
|
// x1, y1 is the width, height of the reference grid |
|
img->x0 = 0; |
|
img->y0 = 0; |
|
img->x1 = (avctx->width - 1) * parameters->subsampling_dx + 1; |
|
img->y1 = (avctx->height - 1) * parameters->subsampling_dy + 1; |
|
|
|
return img; |
|
} |
|
|
|
static av_cold int libopenjpeg_encode_init(AVCodecContext *avctx) |
|
{ |
|
LibOpenJPEGContext *ctx = avctx->priv_data; |
|
int err = AVERROR(ENOMEM); |
|
|
|
opj_set_default_encoder_parameters(&ctx->enc_params); |
|
|
|
ctx->enc_params.cp_rsiz = ctx->profile; |
|
ctx->enc_params.mode = !!avctx->global_quality; |
|
ctx->enc_params.cp_cinema = ctx->cinema_mode; |
|
ctx->enc_params.prog_order = ctx->prog_order; |
|
ctx->enc_params.numresolution = ctx->numresolution; |
|
ctx->enc_params.cp_disto_alloc = ctx->disto_alloc; |
|
ctx->enc_params.cp_fixed_alloc = ctx->fixed_alloc; |
|
ctx->enc_params.cp_fixed_quality = ctx->fixed_quality; |
|
ctx->enc_params.tcp_numlayers = ctx->numlayers; |
|
ctx->enc_params.tcp_rates[0] = FFMAX(avctx->compression_level, 0) * 2; |
|
|
|
if (ctx->cinema_mode > 0) { |
|
cinema_parameters(&ctx->enc_params); |
|
} |
|
|
|
ctx->image = mj2_create_image(avctx, &ctx->enc_params); |
|
if (!ctx->image) { |
|
av_log(avctx, AV_LOG_ERROR, "Error creating the mj2 image\n"); |
|
err = AVERROR(EINVAL); |
|
goto fail; |
|
} |
|
|
|
return 0; |
|
|
|
fail: |
|
opj_image_destroy(ctx->image); |
|
ctx->image = NULL; |
|
return err; |
|
} |
|
|
|
static int libopenjpeg_copy_packed8(AVCodecContext *avctx, const AVFrame *frame, opj_image_t *image) |
|
{ |
|
int compno; |
|
int x; |
|
int y; |
|
int *image_line; |
|
int frame_index; |
|
const int numcomps = image->numcomps; |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
if (image->comps[compno].w > frame->linesize[0] / numcomps) { |
|
av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); |
|
return 0; |
|
} |
|
} |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
for (y = 0; y < avctx->height; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
frame_index = y * frame->linesize[0] + compno; |
|
for (x = 0; x < avctx->width; ++x) { |
|
image_line[x] = frame->data[0][frame_index]; |
|
frame_index += numcomps; |
|
} |
|
for (; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - 1]; |
|
} |
|
} |
|
for (; y < image->comps[compno].h; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
for (x = 0; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - image->comps[compno].w]; |
|
} |
|
} |
|
} |
|
|
|
return 1; |
|
} |
|
|
|
// for XYZ 12 bit |
|
static int libopenjpeg_copy_packed12(AVCodecContext *avctx, const AVFrame *frame, opj_image_t *image) |
|
{ |
|
int compno; |
|
int x, y; |
|
int *image_line; |
|
int frame_index; |
|
const int numcomps = image->numcomps; |
|
uint16_t *frame_ptr = (uint16_t *)frame->data[0]; |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
if (image->comps[compno].w > frame->linesize[0] / numcomps) { |
|
av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); |
|
return 0; |
|
} |
|
} |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
for (y = 0; y < avctx->height; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
frame_index = y * (frame->linesize[0] / 2) + compno; |
|
for (x = 0; x < avctx->width; ++x) { |
|
image_line[x] = frame_ptr[frame_index] >> 4; |
|
frame_index += numcomps; |
|
} |
|
for (; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - 1]; |
|
} |
|
} |
|
for (; y < image->comps[compno].h; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
for (x = 0; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - image->comps[compno].w]; |
|
} |
|
} |
|
} |
|
|
|
return 1; |
|
} |
|
|
|
static int libopenjpeg_copy_packed16(AVCodecContext *avctx, const AVFrame *frame, opj_image_t *image) |
|
{ |
|
int compno; |
|
int x; |
|
int y; |
|
int *image_line; |
|
int frame_index; |
|
const int numcomps = image->numcomps; |
|
uint16_t *frame_ptr = (uint16_t*)frame->data[0]; |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
if (image->comps[compno].w > frame->linesize[0] / numcomps) { |
|
av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); |
|
return 0; |
|
} |
|
} |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
for (y = 0; y < avctx->height; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
frame_index = y * (frame->linesize[0] / 2) + compno; |
|
for (x = 0; x < avctx->width; ++x) { |
|
image_line[x] = frame_ptr[frame_index]; |
|
frame_index += numcomps; |
|
} |
|
for (; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - 1]; |
|
} |
|
} |
|
for (; y < image->comps[compno].h; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
for (x = 0; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - image->comps[compno].w]; |
|
} |
|
} |
|
} |
|
|
|
return 1; |
|
} |
|
|
|
static int libopenjpeg_copy_unpacked8(AVCodecContext *avctx, const AVFrame *frame, opj_image_t *image) |
|
{ |
|
int compno; |
|
int x; |
|
int y; |
|
int width; |
|
int height; |
|
int *image_line; |
|
int frame_index; |
|
const int numcomps = image->numcomps; |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
if (image->comps[compno].w > frame->linesize[compno]) { |
|
av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); |
|
return 0; |
|
} |
|
} |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
width = avctx->width / image->comps[compno].dx; |
|
height = avctx->height / image->comps[compno].dy; |
|
for (y = 0; y < height; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
frame_index = y * frame->linesize[compno]; |
|
for (x = 0; x < width; ++x) |
|
image_line[x] = frame->data[compno][frame_index++]; |
|
for (; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - 1]; |
|
} |
|
} |
|
for (; y < image->comps[compno].h; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
for (x = 0; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - image->comps[compno].w]; |
|
} |
|
} |
|
} |
|
|
|
return 1; |
|
} |
|
|
|
static int libopenjpeg_copy_unpacked16(AVCodecContext *avctx, const AVFrame *frame, opj_image_t *image) |
|
{ |
|
int compno; |
|
int x; |
|
int y; |
|
int width; |
|
int height; |
|
int *image_line; |
|
int frame_index; |
|
const int numcomps = image->numcomps; |
|
uint16_t *frame_ptr; |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
if (image->comps[compno].w > frame->linesize[compno]) { |
|
av_log(avctx, AV_LOG_ERROR, "Error: frame's linesize is too small for the image\n"); |
|
return 0; |
|
} |
|
} |
|
|
|
for (compno = 0; compno < numcomps; ++compno) { |
|
width = avctx->width / image->comps[compno].dx; |
|
height = avctx->height / image->comps[compno].dy; |
|
frame_ptr = (uint16_t *)frame->data[compno]; |
|
for (y = 0; y < height; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
frame_index = y * (frame->linesize[compno] / 2); |
|
for (x = 0; x < width; ++x) |
|
image_line[x] = frame_ptr[frame_index++]; |
|
for (; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - 1]; |
|
} |
|
} |
|
for (; y < image->comps[compno].h; ++y) { |
|
image_line = image->comps[compno].data + y * image->comps[compno].w; |
|
for (x = 0; x < image->comps[compno].w; ++x) { |
|
image_line[x] = image_line[x - image->comps[compno].w]; |
|
} |
|
} |
|
} |
|
|
|
return 1; |
|
} |
|
|
|
static int libopenjpeg_encode_frame(AVCodecContext *avctx, AVPacket *pkt, |
|
const AVFrame *frame, int *got_packet) |
|
{ |
|
LibOpenJPEGContext *ctx = avctx->priv_data; |
|
opj_image_t *image = ctx->image; |
|
opj_cinfo_t *compress = NULL; |
|
opj_cio_t *stream = NULL; |
|
int cpyresult = 0; |
|
int ret, len; |
|
AVFrame *gbrframe; |
|
|
|
switch (avctx->pix_fmt) { |
|
case AV_PIX_FMT_RGB24: |
|
case AV_PIX_FMT_RGBA: |
|
case AV_PIX_FMT_YA8: |
|
cpyresult = libopenjpeg_copy_packed8(avctx, frame, image); |
|
break; |
|
case AV_PIX_FMT_XYZ12: |
|
cpyresult = libopenjpeg_copy_packed12(avctx, frame, image); |
|
break; |
|
case AV_PIX_FMT_RGB48: |
|
case AV_PIX_FMT_RGBA64: |
|
case AV_PIX_FMT_YA16: |
|
cpyresult = libopenjpeg_copy_packed16(avctx, frame, image); |
|
break; |
|
case AV_PIX_FMT_GBR24P: |
|
case AV_PIX_FMT_GBRP9: |
|
case AV_PIX_FMT_GBRP10: |
|
case AV_PIX_FMT_GBRP12: |
|
case AV_PIX_FMT_GBRP14: |
|
case AV_PIX_FMT_GBRP16: |
|
gbrframe = av_frame_clone(frame); |
|
if (!gbrframe) |
|
return AVERROR(ENOMEM); |
|
gbrframe->data[0] = frame->data[2]; // swap to be rgb |
|
gbrframe->data[1] = frame->data[0]; |
|
gbrframe->data[2] = frame->data[1]; |
|
gbrframe->linesize[0] = frame->linesize[2]; |
|
gbrframe->linesize[1] = frame->linesize[0]; |
|
gbrframe->linesize[2] = frame->linesize[1]; |
|
if (avctx->pix_fmt == AV_PIX_FMT_GBR24P) { |
|
cpyresult = libopenjpeg_copy_unpacked8(avctx, gbrframe, image); |
|
} else { |
|
cpyresult = libopenjpeg_copy_unpacked16(avctx, gbrframe, image); |
|
} |
|
av_frame_free(&gbrframe); |
|
break; |
|
case AV_PIX_FMT_GRAY8: |
|
case AV_PIX_FMT_YUV410P: |
|
case AV_PIX_FMT_YUV411P: |
|
case AV_PIX_FMT_YUV420P: |
|
case AV_PIX_FMT_YUV422P: |
|
case AV_PIX_FMT_YUV440P: |
|
case AV_PIX_FMT_YUV444P: |
|
case AV_PIX_FMT_YUVA420P: |
|
case AV_PIX_FMT_YUVA422P: |
|
case AV_PIX_FMT_YUVA444P: |
|
cpyresult = libopenjpeg_copy_unpacked8(avctx, frame, image); |
|
break; |
|
case AV_PIX_FMT_GRAY16: |
|
case AV_PIX_FMT_YUV420P9: |
|
case AV_PIX_FMT_YUV422P9: |
|
case AV_PIX_FMT_YUV444P9: |
|
case AV_PIX_FMT_YUVA420P9: |
|
case AV_PIX_FMT_YUVA422P9: |
|
case AV_PIX_FMT_YUVA444P9: |
|
case AV_PIX_FMT_YUV444P10: |
|
case AV_PIX_FMT_YUV422P10: |
|
case AV_PIX_FMT_YUV420P10: |
|
case AV_PIX_FMT_YUVA444P10: |
|
case AV_PIX_FMT_YUVA422P10: |
|
case AV_PIX_FMT_YUVA420P10: |
|
case AV_PIX_FMT_YUV420P12: |
|
case AV_PIX_FMT_YUV422P12: |
|
case AV_PIX_FMT_YUV444P12: |
|
case AV_PIX_FMT_YUV420P14: |
|
case AV_PIX_FMT_YUV422P14: |
|
case AV_PIX_FMT_YUV444P14: |
|
case AV_PIX_FMT_YUV444P16: |
|
case AV_PIX_FMT_YUV422P16: |
|
case AV_PIX_FMT_YUV420P16: |
|
case AV_PIX_FMT_YUVA444P16: |
|
case AV_PIX_FMT_YUVA422P16: |
|
case AV_PIX_FMT_YUVA420P16: |
|
cpyresult = libopenjpeg_copy_unpacked16(avctx, frame, image); |
|
break; |
|
default: |
|
av_log(avctx, AV_LOG_ERROR, |
|
"The frame's pixel format '%s' is not supported\n", |
|
av_get_pix_fmt_name(avctx->pix_fmt)); |
|
return AVERROR(EINVAL); |
|
break; |
|
} |
|
|
|
if (!cpyresult) { |
|
av_log(avctx, AV_LOG_ERROR, |
|
"Could not copy the frame data to the internal image buffer\n"); |
|
return -1; |
|
} |
|
|
|
compress = opj_create_compress(ctx->format); |
|
if (!compress) { |
|
av_log(avctx, AV_LOG_ERROR, "Error creating the compressor\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
opj_setup_encoder(compress, &ctx->enc_params, image); |
|
|
|
stream = opj_cio_open((opj_common_ptr) compress, NULL, 0); |
|
if (!stream) { |
|
av_log(avctx, AV_LOG_ERROR, "Error creating the cio stream\n"); |
|
return AVERROR(ENOMEM); |
|
} |
|
|
|
memset(&ctx->event_mgr, 0, sizeof(ctx->event_mgr)); |
|
ctx->event_mgr.info_handler = info_callback; |
|
ctx->event_mgr.error_handler = error_callback; |
|
ctx->event_mgr.warning_handler = warning_callback; |
|
opj_set_event_mgr((opj_common_ptr) compress, &ctx->event_mgr, avctx); |
|
|
|
if (!opj_encode(compress, stream, image, NULL)) { |
|
av_log(avctx, AV_LOG_ERROR, "Error during the opj encode\n"); |
|
return -1; |
|
} |
|
|
|
len = cio_tell(stream); |
|
if ((ret = ff_alloc_packet2(avctx, pkt, len)) < 0) { |
|
return ret; |
|
} |
|
|
|
memcpy(pkt->data, stream->buffer, len); |
|
pkt->flags |= AV_PKT_FLAG_KEY; |
|
*got_packet = 1; |
|
|
|
opj_cio_close(stream); |
|
stream = NULL; |
|
opj_destroy_compress(compress); |
|
compress = NULL; |
|
|
|
return 0; |
|
} |
|
|
|
static av_cold int libopenjpeg_encode_close(AVCodecContext *avctx) |
|
{ |
|
LibOpenJPEGContext *ctx = avctx->priv_data; |
|
|
|
opj_image_destroy(ctx->image); |
|
ctx->image = NULL; |
|
return 0; |
|
} |
|
|
|
#define OFFSET(x) offsetof(LibOpenJPEGContext, x) |
|
#define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM |
|
static const AVOption options[] = { |
|
{ "format", "Codec Format", OFFSET(format), AV_OPT_TYPE_INT, { .i64 = CODEC_JP2 }, CODEC_J2K, CODEC_JP2, VE, "format" }, |
|
{ "j2k", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CODEC_J2K }, 0, 0, VE, "format" }, |
|
{ "jp2", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CODEC_JP2 }, 0, 0, VE, "format" }, |
|
{ "profile", NULL, OFFSET(profile), AV_OPT_TYPE_INT, { .i64 = STD_RSIZ }, STD_RSIZ, CINEMA4K, VE, "profile" }, |
|
{ "jpeg2000", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = STD_RSIZ }, 0, 0, VE, "profile" }, |
|
{ "cinema2k", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CINEMA2K }, 0, 0, VE, "profile" }, |
|
{ "cinema4k", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CINEMA4K }, 0, 0, VE, "profile" }, |
|
{ "cinema_mode", "Digital Cinema", OFFSET(cinema_mode), AV_OPT_TYPE_INT, { .i64 = OFF }, OFF, CINEMA4K_24, VE, "cinema_mode" }, |
|
{ "off", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = OFF }, 0, 0, VE, "cinema_mode" }, |
|
{ "2k_24", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CINEMA2K_24 }, 0, 0, VE, "cinema_mode" }, |
|
{ "2k_48", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CINEMA2K_48 }, 0, 0, VE, "cinema_mode" }, |
|
{ "4k_24", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CINEMA4K_24 }, 0, 0, VE, "cinema_mode" }, |
|
{ "prog_order", "Progression Order", OFFSET(prog_order), AV_OPT_TYPE_INT, { .i64 = LRCP }, LRCP, CPRL, VE, "prog_order" }, |
|
{ "lrcp", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = LRCP }, 0, 0, VE, "prog_order" }, |
|
{ "rlcp", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = RLCP }, 0, 0, VE, "prog_order" }, |
|
{ "rpcl", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = RPCL }, 0, 0, VE, "prog_order" }, |
|
{ "pcrl", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = PCRL }, 0, 0, VE, "prog_order" }, |
|
{ "cprl", NULL, 0, AV_OPT_TYPE_CONST, { .i64 = CPRL }, 0, 0, VE, "prog_order" }, |
|
{ "numresolution", NULL, OFFSET(numresolution), AV_OPT_TYPE_INT, { .i64 = 6 }, 1, INT_MAX, VE }, |
|
{ "numlayers", NULL, OFFSET(numlayers), AV_OPT_TYPE_INT, { .i64 = 1 }, 1, 10, VE }, |
|
{ "disto_alloc", NULL, OFFSET(disto_alloc), AV_OPT_TYPE_INT, { .i64 = 1 }, 0, 1, VE }, |
|
{ "fixed_alloc", NULL, OFFSET(fixed_alloc), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 1, VE }, |
|
{ "fixed_quality", NULL, OFFSET(fixed_quality), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 1, VE }, |
|
{ NULL }, |
|
}; |
|
|
|
static const AVClass openjpeg_class = { |
|
.class_name = "libopenjpeg", |
|
.item_name = av_default_item_name, |
|
.option = options, |
|
.version = LIBAVUTIL_VERSION_INT, |
|
}; |
|
|
|
AVCodec ff_libopenjpeg_encoder = { |
|
.name = "libopenjpeg", |
|
.long_name = NULL_IF_CONFIG_SMALL("OpenJPEG JPEG 2000"), |
|
.type = AVMEDIA_TYPE_VIDEO, |
|
.id = AV_CODEC_ID_JPEG2000, |
|
.priv_data_size = sizeof(LibOpenJPEGContext), |
|
.init = libopenjpeg_encode_init, |
|
.encode2 = libopenjpeg_encode_frame, |
|
.close = libopenjpeg_encode_close, |
|
.capabilities = CODEC_CAP_FRAME_THREADS | CODEC_CAP_INTRA_ONLY, |
|
.pix_fmts = (const enum AVPixelFormat[]) { |
|
AV_PIX_FMT_RGB24, AV_PIX_FMT_RGBA, AV_PIX_FMT_RGB48, |
|
AV_PIX_FMT_RGBA64, AV_PIX_FMT_GBR24P, |
|
AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16, |
|
AV_PIX_FMT_GRAY8, AV_PIX_FMT_YA8, AV_PIX_FMT_GRAY16, AV_PIX_FMT_YA16, |
|
AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, |
|
AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUVA422P, |
|
AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUVA444P, |
|
AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, |
|
AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9, |
|
AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, |
|
AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10, |
|
AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, |
|
AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, |
|
AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, |
|
AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16, |
|
AV_PIX_FMT_XYZ12, |
|
AV_PIX_FMT_NONE |
|
}, |
|
.priv_class = &openjpeg_class, |
|
};
|
|
|