From 96dca089b19bf55293136277f5b081c56146e10e Mon Sep 17 00:00:00 2001 From: Anton Khirnov Date: Tue, 21 Jul 2015 18:57:49 +0200 Subject: [PATCH] qsvdec: move qsv_process_data() from qsvdec_h264 to the common code It will be shared with the upcoming mpeg2 and hevc decoders. --- libavcodec/qsvdec.c | 109 +++++++++++++++++++++++++++++++++++++-- libavcodec/qsvdec.h | 12 +++-- libavcodec/qsvdec_h264.c | 107 ++------------------------------------ 3 files changed, 118 insertions(+), 110 deletions(-) diff --git a/libavcodec/qsvdec.c b/libavcodec/qsvdec.c index ef781d9d8e..5263bf5971 100644 --- a/libavcodec/qsvdec.c +++ b/libavcodec/qsvdec.c @@ -34,6 +34,7 @@ #include "avcodec.h" #include "internal.h" +#include "qsv.h" #include "qsv_internal.h" #include "qsvdec.h" @@ -209,9 +210,9 @@ static QSVFrame *find_frame(QSVContext *q, mfxFrameSurface1 *surf) return NULL; } -int ff_qsv_decode(AVCodecContext *avctx, QSVContext *q, - AVFrame *frame, int *got_frame, - AVPacket *avpkt) +static int qsv_decode(AVCodecContext *avctx, QSVContext *q, + AVFrame *frame, int *got_frame, + AVPacket *avpkt) { QSVFrame *out_frame; mfxFrameSurface1 *insurf; @@ -317,8 +318,110 @@ int ff_qsv_decode_close(QSVContext *q) av_fifo_free(q->async_fifo); q->async_fifo = NULL; + av_parser_close(q->parser); + avcodec_free_context(&q->avctx_internal); + if (q->internal_session) MFXClose(q->internal_session); return 0; } + +int ff_qsv_process_data(AVCodecContext *avctx, QSVContext *q, + AVFrame *frame, int *got_frame, AVPacket *pkt) +{ + uint8_t *dummy_data; + int dummy_size; + int ret; + + if (!q->avctx_internal) { + q->avctx_internal = avcodec_alloc_context3(NULL); + if (!q->avctx_internal) + return AVERROR(ENOMEM); + + if (avctx->extradata) { + q->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE); + if (!q->avctx_internal->extradata) + return AVERROR(ENOMEM); + + memcpy(q->avctx_internal->extradata, avctx->extradata, + avctx->extradata_size); + q->avctx_internal->extradata_size = avctx->extradata_size; + } + + q->parser = av_parser_init(avctx->codec_id); + if (!q->parser) + return AVERROR(ENOMEM); + + q->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES; + q->orig_pix_fmt = AV_PIX_FMT_NONE; + } + + if (!pkt->size) + return qsv_decode(avctx, q, frame, got_frame, pkt); + + /* we assume the packets are already split properly and want + * just the codec parameters here */ + av_parser_parse2(q->parser, q->avctx_internal, + &dummy_data, &dummy_size, + pkt->data, pkt->size, pkt->pts, pkt->dts, + pkt->pos); + + /* TODO: flush delayed frames on reinit */ + if (q->parser->format != q->orig_pix_fmt || + q->parser->coded_width != avctx->coded_width || + q->parser->coded_height != avctx->coded_height) { + mfxSession session = NULL; + + enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV, + AV_PIX_FMT_NONE, + AV_PIX_FMT_NONE }; + enum AVPixelFormat qsv_format; + + qsv_format = ff_qsv_map_pixfmt(q->parser->format); + if (qsv_format < 0) { + av_log(avctx, AV_LOG_ERROR, + "Only 8-bit YUV420 streams are supported.\n"); + ret = AVERROR(ENOSYS); + goto reinit_fail; + } + + q->orig_pix_fmt = q->parser->format; + avctx->pix_fmt = pix_fmts[1] = qsv_format; + avctx->width = q->parser->width; + avctx->height = q->parser->height; + avctx->coded_width = q->parser->coded_width; + avctx->coded_height = q->parser->coded_height; + avctx->level = q->avctx_internal->level; + avctx->profile = q->avctx_internal->profile; + + ret = ff_get_format(avctx, pix_fmts); + if (ret < 0) + goto reinit_fail; + + avctx->pix_fmt = ret; + + if (avctx->hwaccel_context) { + AVQSVContext *user_ctx = avctx->hwaccel_context; + session = user_ctx->session; + q->iopattern = user_ctx->iopattern; + q->ext_buffers = user_ctx->ext_buffers; + q->nb_ext_buffers = user_ctx->nb_ext_buffers; + } + + ret = ff_qsv_decode_init(avctx, q, session); + if (ret < 0) + goto reinit_fail; + } + + return qsv_decode(avctx, q, frame, got_frame, pkt); + +reinit_fail: + q->orig_pix_fmt = q->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE; + return ret; +} + +void ff_qsv_decode_flush(AVCodecContext *avctx, QSVContext *q) +{ + q->orig_pix_fmt = AV_PIX_FMT_NONE; +} diff --git a/libavcodec/qsvdec.h b/libavcodec/qsvdec.h index 01f7690137..d5f2bce254 100644 --- a/libavcodec/qsvdec.h +++ b/libavcodec/qsvdec.h @@ -50,6 +50,11 @@ typedef struct QSVContext { AVFifoBuffer *async_fifo; + // the internal parser and codec context for parsing the data + AVCodecParserContext *parser; + AVCodecContext *avctx_internal; + enum AVPixelFormat orig_pix_fmt; + // options set by the caller int async_depth; int iopattern; @@ -62,9 +67,10 @@ int ff_qsv_map_pixfmt(enum AVPixelFormat format); int ff_qsv_decode_init(AVCodecContext *s, QSVContext *q, mfxSession session); -int ff_qsv_decode(AVCodecContext *s, QSVContext *q, - AVFrame *frame, int *got_frame, - AVPacket *avpkt); +int ff_qsv_process_data(AVCodecContext *avctx, QSVContext *q, + AVFrame *frame, int *got_frame, AVPacket *pkt); + +void ff_qsv_decode_flush(AVCodecContext *avctx, QSVContext *q); int ff_qsv_decode_close(QSVContext *q); diff --git a/libavcodec/qsvdec_h264.c b/libavcodec/qsvdec_h264.c index b6527f0cc0..a9f48e8b43 100644 --- a/libavcodec/qsvdec_h264.c +++ b/libavcodec/qsvdec_h264.c @@ -41,11 +41,6 @@ typedef struct QSVH264Context { AVClass *class; QSVContext qsv; - // the internal parser and codec context for parsing the data - AVCodecParserContext *parser; - AVCodecContext *avctx_internal; - enum AVPixelFormat orig_pix_fmt; - // the filter for converting to Annex B AVBitStreamFilterContext *bsf; @@ -81,8 +76,6 @@ static av_cold int qsv_decode_close(AVCodecContext *avctx) av_fifo_free(s->packet_fifo); av_bitstream_filter_close(s->bsf); - av_parser_close(s->parser); - avcodec_free_context(&s->avctx_internal); return 0; } @@ -92,7 +85,6 @@ static av_cold int qsv_decode_init(AVCodecContext *avctx) QSVH264Context *s = avctx->priv_data; int ret; - s->orig_pix_fmt = AV_PIX_FMT_NONE; s->packet_fifo = av_fifo_alloc(sizeof(AVPacket)); if (!s->packet_fifo) { @@ -106,30 +98,6 @@ static av_cold int qsv_decode_init(AVCodecContext *avctx) goto fail; } - s->avctx_internal = avcodec_alloc_context3(NULL); - if (!s->avctx_internal) { - ret = AVERROR(ENOMEM); - goto fail; - } - - if (avctx->extradata) { - s->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE); - if (!s->avctx_internal->extradata) { - ret = AVERROR(ENOMEM); - goto fail; - } - memcpy(s->avctx_internal->extradata, avctx->extradata, - avctx->extradata_size); - s->avctx_internal->extradata_size = avctx->extradata_size; - } - - s->parser = av_parser_init(AV_CODEC_ID_H264); - if (!s->parser) { - ret = AVERROR(ENOMEM); - goto fail; - } - s->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES; - s->qsv.iopattern = MFX_IOPATTERN_OUT_SYSTEM_MEMORY; return 0; @@ -138,75 +106,6 @@ fail: return ret; } -static int qsv_process_data(AVCodecContext *avctx, AVFrame *frame, - int *got_frame, AVPacket *pkt) -{ - QSVH264Context *s = avctx->priv_data; - uint8_t *dummy_data; - int dummy_size; - int ret; - - /* we assume the packets are already split properly and want - * just the codec parameters here */ - av_parser_parse2(s->parser, s->avctx_internal, - &dummy_data, &dummy_size, - pkt->data, pkt->size, pkt->pts, pkt->dts, - pkt->pos); - - /* TODO: flush delayed frames on reinit */ - if (s->parser->format != s->orig_pix_fmt || - s->parser->coded_width != avctx->coded_width || - s->parser->coded_height != avctx->coded_height) { - mfxSession session = NULL; - - enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV, - AV_PIX_FMT_NONE, - AV_PIX_FMT_NONE }; - enum AVPixelFormat qsv_format; - - qsv_format = ff_qsv_map_pixfmt(s->parser->format); - if (qsv_format < 0) { - av_log(avctx, AV_LOG_ERROR, - "Only 8-bit YUV420 streams are supported.\n"); - ret = AVERROR(ENOSYS); - goto reinit_fail; - } - - s->orig_pix_fmt = s->parser->format; - avctx->pix_fmt = pix_fmts[1] = qsv_format; - avctx->width = s->parser->width; - avctx->height = s->parser->height; - avctx->coded_width = s->parser->coded_width; - avctx->coded_height = s->parser->coded_height; - avctx->level = s->avctx_internal->level; - avctx->profile = s->avctx_internal->profile; - - ret = ff_get_format(avctx, pix_fmts); - if (ret < 0) - goto reinit_fail; - - avctx->pix_fmt = ret; - - if (avctx->hwaccel_context) { - AVQSVContext *user_ctx = avctx->hwaccel_context; - session = user_ctx->session; - s->qsv.iopattern = user_ctx->iopattern; - s->qsv.ext_buffers = user_ctx->ext_buffers; - s->qsv.nb_ext_buffers = user_ctx->nb_ext_buffers; - } - - ret = ff_qsv_decode_init(avctx, &s->qsv, session); - if (ret < 0) - goto reinit_fail; - } - - return ff_qsv_decode(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered); - -reinit_fail: - s->orig_pix_fmt = s->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE; - return ret; -} - static int qsv_decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt) { @@ -239,7 +138,7 @@ static int qsv_decode_frame(AVCodecContext *avctx, void *data, /* no more data */ if (av_fifo_size(s->packet_fifo) < sizeof(AVPacket)) - return avpkt->size ? avpkt->size : ff_qsv_decode(avctx, &s->qsv, frame, got_frame, avpkt); + return avpkt->size ? avpkt->size : ff_qsv_process_data(avctx, &s->qsv, frame, got_frame, avpkt); if (s->filtered_data != s->input_ref.data) av_freep(&s->filtered_data); @@ -259,7 +158,7 @@ static int qsv_decode_frame(AVCodecContext *avctx, void *data, s->pkt_filtered.size = size; } - ret = qsv_process_data(avctx, frame, got_frame, &s->pkt_filtered); + ret = ff_qsv_process_data(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered); if (ret < 0) return ret; @@ -275,7 +174,7 @@ static void qsv_decode_flush(AVCodecContext *avctx) QSVH264Context *s = avctx->priv_data; qsv_clear_buffers(s); - s->orig_pix_fmt = AV_PIX_FMT_NONE; + ff_qsv_decode_flush(avctx, &s->qsv); } AVHWAccel ff_h264_qsv_hwaccel = {