|
|
|
/*
|
|
|
|
* Image format
|
|
|
|
* Copyright (c) 2000, 2001, 2002 Fabrice Bellard
|
|
|
|
* Copyright (c) 2004 Michael Niedermayer
|
|
|
|
*
|
|
|
|
* This file is part of FFmpeg.
|
|
|
|
*
|
|
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define _DEFAULT_SOURCE
|
|
|
|
#define _BSD_SOURCE
|
|
|
|
#include <sys/stat.h>
|
|
|
|
#include "libavutil/avstring.h"
|
|
|
|
#include "libavutil/log.h"
|
|
|
|
#include "libavutil/opt.h"
|
|
|
|
#include "libavutil/pixdesc.h"
|
|
|
|
#include "libavutil/parseutils.h"
|
|
|
|
#include "libavutil/intreadwrite.h"
|
|
|
|
#include "libavcodec/gif.h"
|
|
|
|
#include "avformat.h"
|
|
|
|
#include "avio_internal.h"
|
|
|
|
#include "internal.h"
|
|
|
|
#include "img2.h"
|
|
|
|
#include "libavcodec/mjpeg.h"
|
|
|
|
#include "libavcodec/xwd.h"
|
|
|
|
#include "subtitles.h"
|
|
|
|
|
|
|
|
#if HAVE_GLOB
|
|
|
|
/* Locally define as 0 (bitwise-OR no-op) any missing glob options that
|
|
|
|
are non-posix glibc/bsd extensions. */
|
|
|
|
#ifndef GLOB_NOMAGIC
|
|
|
|
#define GLOB_NOMAGIC 0
|
|
|
|
#endif
|
|
|
|
#ifndef GLOB_BRACE
|
|
|
|
#define GLOB_BRACE 0
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif /* HAVE_GLOB */
|
|
|
|
|
|
|
|
static const int sizes[][2] = {
|
|
|
|
{ 640, 480 },
|
|
|
|
{ 720, 480 },
|
|
|
|
{ 720, 576 },
|
|
|
|
{ 352, 288 },
|
|
|
|
{ 352, 240 },
|
|
|
|
{ 160, 128 },
|
|
|
|
{ 512, 384 },
|
|
|
|
{ 640, 352 },
|
|
|
|
{ 640, 240 },
|
|
|
|
};
|
|
|
|
|
|
|
|
static int infer_size(int *width_ptr, int *height_ptr, int size)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < FF_ARRAY_ELEMS(sizes); i++) {
|
|
|
|
if ((sizes[i][0] * sizes[i][1]) == size) {
|
|
|
|
*width_ptr = sizes[i][0];
|
|
|
|
*height_ptr = sizes[i][1];
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int is_glob(const char *path)
|
|
|
|
{
|
|
|
|
#if HAVE_GLOB
|
|
|
|
size_t span = 0;
|
|
|
|
const char *p = path;
|
|
|
|
|
|
|
|
while (p = strchr(p, '%')) {
|
|
|
|
if (*(++p) == '%') {
|
|
|
|
++p;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (span = strspn(p, "*?[]{}"))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
/* Did we hit a glob char or get to the end? */
|
|
|
|
return span != 0;
|
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Get index range of image files matched by path.
|
|
|
|
*
|
|
|
|
* @param pfirst_index pointer to index updated with the first number in the range
|
|
|
|
* @param plast_index pointer to index updated with the last number in the range
|
|
|
|
* @param path path which has to be matched by the image files in the range
|
|
|
|
* @param start_index minimum accepted value for the first index in the range
|
|
|
|
* @return -1 if no image file could be found
|
|
|
|
*/
|
|
|
|
static int find_image_range(AVIOContext *pb, int *pfirst_index, int *plast_index,
|
|
|
|
const char *path, int start_index, int start_index_range)
|
|
|
|
{
|
|
|
|
char buf[1024];
|
|
|
|
int range, last_index, range1, first_index;
|
|
|
|
|
|
|
|
/* find the first image */
|
|
|
|
for (first_index = start_index; first_index < start_index + start_index_range; first_index++) {
|
|
|
|
if (av_get_frame_filename(buf, sizeof(buf), path, first_index) < 0) {
|
|
|
|
*pfirst_index =
|
|
|
|
*plast_index = 1;
|
|
|
|
if (pb || avio_check(buf, AVIO_FLAG_READ) > 0)
|
|
|
|
return 0;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
if (avio_check(buf, AVIO_FLAG_READ) > 0)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (first_index == start_index + start_index_range)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
/* find the last image */
|
|
|
|
last_index = first_index;
|
|
|
|
for (;;) {
|
|
|
|
range = 0;
|
|
|
|
for (;;) {
|
|
|
|
if (!range)
|
|
|
|
range1 = 1;
|
|
|
|
else
|
|
|
|
range1 = 2 * range;
|
|
|
|
if (av_get_frame_filename(buf, sizeof(buf), path,
|
|
|
|
last_index + range1) < 0)
|
|
|
|
goto fail;
|
|
|
|
if (avio_check(buf, AVIO_FLAG_READ) <= 0)
|
|
|
|
break;
|
|
|
|
range = range1;
|
|
|
|
/* just in case... */
|
|
|
|
if (range >= (1 << 30))
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
/* we are sure than image last_index + range exists */
|
|
|
|
if (!range)
|
|
|
|
break;
|
|
|
|
last_index += range;
|
|
|
|
}
|
|
|
|
*pfirst_index = first_index;
|
|
|
|
*plast_index = last_index;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int img_read_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
if (p->filename && ff_guess_image2_codec(p->filename)) {
|
|
|
|
if (av_filename_number_test(p->filename))
|
|
|
|
return AVPROBE_SCORE_MAX;
|
|
|
|
else if (is_glob(p->filename))
|
|
|
|
return AVPROBE_SCORE_MAX;
|
|
|
|
else if (p->filename[strcspn(p->filename, "*?{")]) // probably PT_GLOB
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 2; // score chosen to be a tad above the image pipes
|
|
|
|
else if (p->buf_size == 0)
|
|
|
|
return 0;
|
|
|
|
else if (av_match_ext(p->filename, "raw") || av_match_ext(p->filename, "gif"))
|
|
|
|
return 5;
|
|
|
|
else
|
|
|
|
return AVPROBE_SCORE_EXTENSION;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int ff_img_read_header(AVFormatContext *s1)
|
|
|
|
{
|
|
|
|
VideoDemuxData *s = s1->priv_data;
|
|
|
|
int first_index = 1, last_index = 1;
|
|
|
|
AVStream *st;
|
|
|
|
enum AVPixelFormat pix_fmt = AV_PIX_FMT_NONE;
|
|
|
|
|
|
|
|
s1->ctx_flags |= AVFMTCTX_NOHEADER;
|
|
|
|
|
|
|
|
st = avformat_new_stream(s1, NULL);
|
|
|
|
if (!st) {
|
|
|
|
return AVERROR(ENOMEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->pixel_format &&
|
|
|
|
(pix_fmt = av_get_pix_fmt(s->pixel_format)) == AV_PIX_FMT_NONE) {
|
|
|
|
av_log(s1, AV_LOG_ERROR, "No such pixel format: %s.\n",
|
|
|
|
s->pixel_format);
|
|
|
|
return AVERROR(EINVAL);
|
|
|
|
}
|
|
|
|
|
|
|
|
av_strlcpy(s->path, s1->url, sizeof(s->path));
|
|
|
|
s->img_number = 0;
|
|
|
|
s->img_count = 0;
|
|
|
|
|
|
|
|
/* find format */
|
|
|
|
if (s1->iformat->flags & AVFMT_NOFILE)
|
|
|
|
s->is_pipe = 0;
|
|
|
|
else {
|
|
|
|
s->is_pipe = 1;
|
|
|
|
st->need_parsing = AVSTREAM_PARSE_FULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->ts_from_file == 2) {
|
|
|
|
#if !HAVE_STRUCT_STAT_ST_MTIM_TV_NSEC
|
|
|
|
av_log(s1, AV_LOG_ERROR, "POSIX.1-2008 not supported, nanosecond file timestamps unavailable\n");
|
|
|
|
return AVERROR(ENOSYS);
|
|
|
|
#endif
|
|
|
|
avpriv_set_pts_info(st, 64, 1, 1000000000);
|
|
|
|
} else if (s->ts_from_file)
|
|
|
|
avpriv_set_pts_info(st, 64, 1, 1);
|
|
|
|
else
|
|
|
|
avpriv_set_pts_info(st, 64, s->framerate.den, s->framerate.num);
|
|
|
|
|
|
|
|
if (s->width && s->height) {
|
|
|
|
st->codecpar->width = s->width;
|
|
|
|
st->codecpar->height = s->height;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!s->is_pipe) {
|
|
|
|
if (s->pattern_type == PT_DEFAULT) {
|
|
|
|
if (s1->pb) {
|
|
|
|
s->pattern_type = PT_NONE;
|
|
|
|
} else
|
|
|
|
s->pattern_type = PT_GLOB_SEQUENCE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->pattern_type == PT_GLOB_SEQUENCE) {
|
|
|
|
s->use_glob = is_glob(s->path);
|
|
|
|
if (s->use_glob) {
|
|
|
|
#if HAVE_GLOB
|
|
|
|
char *p = s->path, *q, *dup;
|
|
|
|
int gerr;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
av_log(s1, AV_LOG_WARNING, "Pattern type 'glob_sequence' is deprecated: "
|
|
|
|
"use pattern_type 'glob' instead\n");
|
|
|
|
#if HAVE_GLOB
|
|
|
|
dup = q = av_strdup(p);
|
|
|
|
while (*q) {
|
|
|
|
/* Do we have room for the next char and a \ insertion? */
|
|
|
|
if ((p - s->path) >= (sizeof(s->path) - 2))
|
|
|
|
break;
|
|
|
|
if (*q == '%' && strspn(q + 1, "%*?[]{}"))
|
|
|
|
++q;
|
|
|
|
else if (strspn(q, "\\*?[]{}"))
|
|
|
|
*p++ = '\\';
|
|
|
|
*p++ = *q++;
|
|
|
|
}
|
|
|
|
*p = 0;
|
|
|
|
av_free(dup);
|
|
|
|
|
|
|
|
gerr = glob(s->path, GLOB_NOCHECK|GLOB_BRACE|GLOB_NOMAGIC, NULL, &s->globstate);
|
|
|
|
if (gerr != 0) {
|
|
|
|
return AVERROR(ENOENT);
|
|
|
|
}
|
|
|
|
first_index = 0;
|
|
|
|
last_index = s->globstate.gl_pathc - 1;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if ((s->pattern_type == PT_GLOB_SEQUENCE && !s->use_glob) || s->pattern_type == PT_SEQUENCE) {
|
|
|
|
if (find_image_range(s1->pb, &first_index, &last_index, s->path,
|
|
|
|
s->start_number, s->start_number_range) < 0) {
|
|
|
|
av_log(s1, AV_LOG_ERROR,
|
|
|
|
"Could find no file with path '%s' and index in the range %d-%d\n",
|
|
|
|
s->path, s->start_number, s->start_number + s->start_number_range - 1);
|
|
|
|
return AVERROR(ENOENT);
|
|
|
|
}
|
|
|
|
} else if (s->pattern_type == PT_GLOB) {
|
|
|
|
#if HAVE_GLOB
|
|
|
|
int gerr;
|
|
|
|
gerr = glob(s->path, GLOB_NOCHECK|GLOB_BRACE|GLOB_NOMAGIC, NULL, &s->globstate);
|
|
|
|
if (gerr != 0) {
|
|
|
|
return AVERROR(ENOENT);
|
|
|
|
}
|
|
|
|
first_index = 0;
|
|
|
|
last_index = s->globstate.gl_pathc - 1;
|
|
|
|
s->use_glob = 1;
|
|
|
|
#else
|
|
|
|
av_log(s1, AV_LOG_ERROR,
|
|
|
|
"Pattern type 'glob' was selected but globbing "
|
|
|
|
"is not supported by this libavformat build\n");
|
|
|
|
return AVERROR(ENOSYS);
|
|
|
|
#endif
|
|
|
|
} else if (s->pattern_type != PT_GLOB_SEQUENCE && s->pattern_type != PT_NONE) {
|
|
|
|
av_log(s1, AV_LOG_ERROR,
|
|
|
|
"Unknown value '%d' for pattern_type option\n", s->pattern_type);
|
|
|
|
return AVERROR(EINVAL);
|
|
|
|
}
|
|
|
|
s->img_first = first_index;
|
|
|
|
s->img_last = last_index;
|
|
|
|
s->img_number = first_index;
|
|
|
|
/* compute duration */
|
|
|
|
if (!s->ts_from_file) {
|
|
|
|
st->start_time = 0;
|
|
|
|
st->duration = last_index - first_index + 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s1->video_codec_id) {
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
|
|
st->codecpar->codec_id = s1->video_codec_id;
|
|
|
|
} else if (s1->audio_codec_id) {
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
st->codecpar->codec_type = AVMEDIA_TYPE_AUDIO;
|
|
|
|
st->codecpar->codec_id = s1->audio_codec_id;
|
|
|
|
} else if (s1->iformat->raw_codec_id) {
|
|
|
|
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
|
|
st->codecpar->codec_id = s1->iformat->raw_codec_id;
|
|
|
|
} else {
|
|
|
|
const char *str = strrchr(s->path, '.');
|
|
|
|
s->split_planes = str && !av_strcasecmp(str + 1, "y");
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
st->codecpar->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
|
|
if (s1->pb) {
|
|
|
|
int probe_buffer_size = 2048;
|
|
|
|
uint8_t *probe_buffer = av_realloc(NULL, probe_buffer_size + AVPROBE_PADDING_SIZE);
|
|
|
|
const AVInputFormat *fmt = NULL;
|
|
|
|
void *fmt_iter = NULL;
|
|
|
|
AVProbeData pd = { 0 };
|
|
|
|
|
|
|
|
if (!probe_buffer)
|
|
|
|
return AVERROR(ENOMEM);
|
|
|
|
|
|
|
|
probe_buffer_size = avio_read(s1->pb, probe_buffer, probe_buffer_size);
|
|
|
|
if (probe_buffer_size < 0) {
|
|
|
|
av_free(probe_buffer);
|
|
|
|
return probe_buffer_size;
|
|
|
|
}
|
|
|
|
memset(probe_buffer + probe_buffer_size, 0, AVPROBE_PADDING_SIZE);
|
|
|
|
|
|
|
|
pd.buf = probe_buffer;
|
|
|
|
pd.buf_size = probe_buffer_size;
|
|
|
|
pd.filename = s1->url;
|
|
|
|
|
|
|
|
while ((fmt = av_demuxer_iterate(&fmt_iter))) {
|
|
|
|
if (fmt->read_header != ff_img_read_header ||
|
|
|
|
!fmt->read_probe ||
|
|
|
|
(fmt->flags & AVFMT_NOFILE) ||
|
|
|
|
!fmt->raw_codec_id)
|
|
|
|
continue;
|
|
|
|
if (fmt->read_probe(&pd) > 0) {
|
|
|
|
st->codecpar->codec_id = fmt->raw_codec_id;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (s1->flags & AVFMT_FLAG_CUSTOM_IO) {
|
|
|
|
avio_seek(s1->pb, 0, SEEK_SET);
|
|
|
|
av_freep(&probe_buffer);
|
|
|
|
} else
|
|
|
|
ffio_rewind_with_probe_data(s1->pb, &probe_buffer, probe_buffer_size);
|
|
|
|
}
|
|
|
|
if (st->codecpar->codec_id == AV_CODEC_ID_NONE)
|
|
|
|
st->codecpar->codec_id = ff_guess_image2_codec(s->path);
|
|
|
|
if (st->codecpar->codec_id == AV_CODEC_ID_LJPEG)
|
|
|
|
st->codecpar->codec_id = AV_CODEC_ID_MJPEG;
|
|
|
|
if (st->codecpar->codec_id == AV_CODEC_ID_ALIAS_PIX) // we cannot distingiush this from BRENDER_PIX
|
|
|
|
st->codecpar->codec_id = AV_CODEC_ID_NONE;
|
|
|
|
}
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO &&
|
|
|
|
pix_fmt != AV_PIX_FMT_NONE)
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
st->codecpar->format = pix_fmt;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int ff_img_read_packet(AVFormatContext *s1, AVPacket *pkt)
|
|
|
|
{
|
|
|
|
VideoDemuxData *s = s1->priv_data;
|
|
|
|
char filename_bytes[1024];
|
|
|
|
char *filename = filename_bytes;
|
|
|
|
int i, res;
|
|
|
|
int size[3] = { 0 }, ret[3] = { 0 };
|
|
|
|
AVIOContext *f[3] = { NULL };
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
AVCodecParameters *par = s1->streams[0]->codecpar;
|
|
|
|
|
|
|
|
if (!s->is_pipe) {
|
|
|
|
/* loop over input */
|
|
|
|
if (s->loop && s->img_number > s->img_last) {
|
|
|
|
s->img_number = s->img_first;
|
|
|
|
}
|
|
|
|
if (s->img_number > s->img_last)
|
|
|
|
return AVERROR_EOF;
|
|
|
|
if (s->pattern_type == PT_NONE) {
|
|
|
|
av_strlcpy(filename_bytes, s->path, sizeof(filename_bytes));
|
|
|
|
} else if (s->use_glob) {
|
|
|
|
#if HAVE_GLOB
|
|
|
|
filename = s->globstate.gl_pathv[s->img_number];
|
|
|
|
#endif
|
|
|
|
} else {
|
|
|
|
if (av_get_frame_filename(filename_bytes, sizeof(filename_bytes),
|
|
|
|
s->path,
|
|
|
|
s->img_number) < 0 && s->img_number > 1)
|
|
|
|
return AVERROR(EIO);
|
|
|
|
}
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
|
|
if (s1->pb &&
|
|
|
|
!strcmp(filename_bytes, s->path) &&
|
|
|
|
!s->loop &&
|
|
|
|
!s->split_planes) {
|
|
|
|
f[i] = s1->pb;
|
|
|
|
} else if (s1->io_open(s1, &f[i], filename, AVIO_FLAG_READ, NULL) < 0) {
|
|
|
|
if (i >= 1)
|
|
|
|
break;
|
|
|
|
av_log(s1, AV_LOG_ERROR, "Could not open file : %s\n",
|
|
|
|
filename);
|
|
|
|
return AVERROR(EIO);
|
|
|
|
}
|
|
|
|
size[i] = avio_size(f[i]);
|
|
|
|
|
|
|
|
if (!s->split_planes)
|
|
|
|
break;
|
|
|
|
filename[strlen(filename) - 1] = 'U' + i;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (par->codec_id == AV_CODEC_ID_NONE) {
|
|
|
|
AVProbeData pd = { 0 };
|
|
|
|
const AVInputFormat *ifmt;
|
|
|
|
uint8_t header[PROBE_BUF_MIN + AVPROBE_PADDING_SIZE];
|
|
|
|
int ret;
|
|
|
|
int score = 0;
|
|
|
|
|
|
|
|
ret = avio_read(f[0], header, PROBE_BUF_MIN);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
memset(header + ret, 0, sizeof(header) - ret);
|
|
|
|
avio_skip(f[0], -ret);
|
|
|
|
pd.buf = header;
|
|
|
|
pd.buf_size = ret;
|
|
|
|
pd.filename = filename;
|
|
|
|
|
|
|
|
ifmt = av_probe_input_format3(&pd, 1, &score);
|
|
|
|
if (ifmt && ifmt->read_packet == ff_img_read_packet && ifmt->raw_codec_id)
|
|
|
|
par->codec_id = ifmt->raw_codec_id;
|
|
|
|
}
|
|
|
|
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
if (par->codec_id == AV_CODEC_ID_RAWVIDEO && !par->width)
|
|
|
|
infer_size(&par->width, &par->height, size[0]);
|
|
|
|
} else {
|
|
|
|
f[0] = s1->pb;
|
|
|
|
if (avio_feof(f[0]) && s->loop && s->is_pipe)
|
|
|
|
avio_seek(f[0], 0, SEEK_SET);
|
|
|
|
if (avio_feof(f[0]))
|
|
|
|
return AVERROR_EOF;
|
|
|
|
if (s->frame_size > 0) {
|
|
|
|
size[0] = s->frame_size;
|
|
|
|
} else if (!s1->streams[0]->parser) {
|
|
|
|
size[0] = avio_size(s1->pb);
|
|
|
|
} else {
|
|
|
|
size[0] = 4096;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
res = av_new_packet(pkt, size[0] + size[1] + size[2]);
|
|
|
|
if (res < 0) {
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
pkt->stream_index = 0;
|
|
|
|
pkt->flags |= AV_PKT_FLAG_KEY;
|
|
|
|
if (s->ts_from_file) {
|
|
|
|
struct stat img_stat;
|
|
|
|
if (stat(filename, &img_stat)) {
|
|
|
|
res = AVERROR(EIO);
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
pkt->pts = (int64_t)img_stat.st_mtime;
|
|
|
|
#if HAVE_STRUCT_STAT_ST_MTIM_TV_NSEC
|
|
|
|
if (s->ts_from_file == 2)
|
|
|
|
pkt->pts = 1000000000*pkt->pts + img_stat.st_mtim.tv_nsec;
|
|
|
|
#endif
|
|
|
|
av_add_index_entry(s1->streams[0], s->img_number, pkt->pts, 0, 0, AVINDEX_KEYFRAME);
|
|
|
|
} else if (!s->is_pipe) {
|
|
|
|
pkt->pts = s->pts;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (s->is_pipe)
|
|
|
|
pkt->pos = avio_tell(f[0]);
|
|
|
|
|
|
|
|
pkt->size = 0;
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
|
|
if (f[i]) {
|
|
|
|
ret[i] = avio_read(f[i], pkt->data + pkt->size, size[i]);
|
|
|
|
if (s->loop && s->is_pipe && ret[i] == AVERROR_EOF) {
|
|
|
|
if (avio_seek(f[i], 0, SEEK_SET) >= 0) {
|
|
|
|
pkt->pos = 0;
|
|
|
|
ret[i] = avio_read(f[i], pkt->data + pkt->size, size[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (!s->is_pipe && f[i] != s1->pb)
|
|
|
|
ff_format_io_close(s1, &f[i]);
|
|
|
|
if (ret[i] > 0)
|
|
|
|
pkt->size += ret[i];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret[0] <= 0 || ret[1] < 0 || ret[2] < 0) {
|
|
|
|
av_packet_unref(pkt);
|
|
|
|
if (ret[0] < 0) {
|
|
|
|
res = ret[0];
|
|
|
|
} else if (ret[1] < 0) {
|
|
|
|
res = ret[1];
|
|
|
|
} else if (ret[2] < 0) {
|
|
|
|
res = ret[2];
|
|
|
|
} else {
|
|
|
|
res = AVERROR_EOF;
|
|
|
|
}
|
|
|
|
goto fail;
|
|
|
|
} else {
|
|
|
|
s->img_count++;
|
|
|
|
s->img_number++;
|
|
|
|
s->pts++;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
fail:
|
|
|
|
if (!s->is_pipe) {
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
|
|
if (f[i] != s1->pb)
|
|
|
|
ff_format_io_close(s1, &f[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int img_read_close(struct AVFormatContext* s1)
|
|
|
|
{
|
|
|
|
#if HAVE_GLOB
|
|
|
|
VideoDemuxData *s = s1->priv_data;
|
|
|
|
if (s->use_glob) {
|
|
|
|
globfree(&s->globstate);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int img_read_seek(AVFormatContext *s, int stream_index, int64_t timestamp, int flags)
|
|
|
|
{
|
|
|
|
VideoDemuxData *s1 = s->priv_data;
|
|
|
|
AVStream *st = s->streams[0];
|
|
|
|
|
|
|
|
if (s1->ts_from_file) {
|
|
|
|
int index = av_index_search_timestamp(st, timestamp, flags);
|
|
|
|
if(index < 0)
|
|
|
|
return -1;
|
|
|
|
s1->img_number = st->index_entries[index].pos;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (timestamp < 0 || !s1->loop && timestamp > s1->img_last - s1->img_first)
|
|
|
|
return -1;
|
|
|
|
s1->img_number = timestamp%(s1->img_last - s1->img_first + 1) + s1->img_first;
|
|
|
|
s1->pts = timestamp;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define OFFSET(x) offsetof(VideoDemuxData, x)
|
|
|
|
#define DEC AV_OPT_FLAG_DECODING_PARAM
|
|
|
|
#define COMMON_OPTIONS \
|
|
|
|
{ "framerate", "set the video framerate", OFFSET(framerate), AV_OPT_TYPE_VIDEO_RATE, {.str = "25"}, 0, INT_MAX, DEC }, \
|
|
|
|
{ "pixel_format", "set video pixel format", OFFSET(pixel_format), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, DEC }, \
|
|
|
|
{ "video_size", "set video size", OFFSET(width), AV_OPT_TYPE_IMAGE_SIZE, {.str = NULL}, 0, 0, DEC }, \
|
|
|
|
{ "loop", "force loop over input file sequence", OFFSET(loop), AV_OPT_TYPE_BOOL, {.i64 = 0 }, 0, 1, DEC }, \
|
|
|
|
{ NULL },
|
|
|
|
|
|
|
|
#if CONFIG_IMAGE2_DEMUXER
|
|
|
|
const AVOption ff_img_options[] = {
|
|
|
|
{ "pattern_type", "set pattern type", OFFSET(pattern_type), AV_OPT_TYPE_INT, {.i64=PT_DEFAULT}, 0, INT_MAX, DEC, "pattern_type"},
|
|
|
|
{ "glob_sequence","select glob/sequence pattern type", 0, AV_OPT_TYPE_CONST, {.i64=PT_GLOB_SEQUENCE}, INT_MIN, INT_MAX, DEC, "pattern_type" },
|
|
|
|
{ "glob", "select glob pattern type", 0, AV_OPT_TYPE_CONST, {.i64=PT_GLOB }, INT_MIN, INT_MAX, DEC, "pattern_type" },
|
|
|
|
{ "sequence", "select sequence pattern type", 0, AV_OPT_TYPE_CONST, {.i64=PT_SEQUENCE }, INT_MIN, INT_MAX, DEC, "pattern_type" },
|
|
|
|
{ "none", "disable pattern matching", 0, AV_OPT_TYPE_CONST, {.i64=PT_NONE }, INT_MIN, INT_MAX, DEC, "pattern_type" },
|
|
|
|
{ "start_number", "set first number in the sequence", OFFSET(start_number), AV_OPT_TYPE_INT, {.i64 = 0 }, INT_MIN, INT_MAX, DEC },
|
|
|
|
{ "start_number_range", "set range for looking at the first sequence number", OFFSET(start_number_range), AV_OPT_TYPE_INT, {.i64 = 5}, 1, INT_MAX, DEC },
|
|
|
|
{ "ts_from_file", "set frame timestamp from file's one", OFFSET(ts_from_file), AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 2, DEC, "ts_type" },
|
|
|
|
{ "none", "none", 0, AV_OPT_TYPE_CONST, {.i64 = 0 }, 0, 2, DEC, "ts_type" },
|
|
|
|
{ "sec", "second precision", 0, AV_OPT_TYPE_CONST, {.i64 = 1 }, 0, 2, DEC, "ts_type" },
|
|
|
|
{ "ns", "nano second precision", 0, AV_OPT_TYPE_CONST, {.i64 = 2 }, 0, 2, DEC, "ts_type" },
|
|
|
|
COMMON_OPTIONS
|
|
|
|
};
|
|
|
|
|
|
|
|
static const AVClass img2_class = {
|
|
|
|
.class_name = "image2 demuxer",
|
|
|
|
.item_name = av_default_item_name,
|
|
|
|
.option = ff_img_options,
|
|
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
|
|
};
|
|
|
|
AVInputFormat ff_image2_demuxer = {
|
|
|
|
.name = "image2",
|
|
|
|
.long_name = NULL_IF_CONFIG_SMALL("image2 sequence"),
|
|
|
|
.priv_data_size = sizeof(VideoDemuxData),
|
|
|
|
.read_probe = img_read_probe,
|
|
|
|
.read_header = ff_img_read_header,
|
|
|
|
.read_packet = ff_img_read_packet,
|
|
|
|
.read_close = img_read_close,
|
|
|
|
.read_seek = img_read_seek,
|
|
|
|
.flags = AVFMT_NOFILE,
|
|
|
|
.priv_class = &img2_class,
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
const AVOption ff_img2pipe_options[] = {
|
|
|
|
{ "frame_size", "force frame size in bytes", OFFSET(frame_size), AV_OPT_TYPE_INT, {.i64 = 0 }, 0, INT_MAX, DEC },
|
|
|
|
COMMON_OPTIONS
|
|
|
|
};
|
|
|
|
|
|
|
|
#if CONFIG_IMAGE2PIPE_DEMUXER
|
|
|
|
static const AVClass img2pipe_class = {
|
|
|
|
.class_name = "image2pipe demuxer",
|
|
|
|
.item_name = av_default_item_name,
|
|
|
|
.option = ff_img2pipe_options,
|
|
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
|
|
};
|
|
|
|
AVInputFormat ff_image2pipe_demuxer = {
|
|
|
|
.name = "image2pipe",
|
|
|
|
.long_name = NULL_IF_CONFIG_SMALL("piped image2 sequence"),
|
|
|
|
.priv_data_size = sizeof(VideoDemuxData),
|
|
|
|
.read_header = ff_img_read_header,
|
|
|
|
.read_packet = ff_img_read_packet,
|
|
|
|
.priv_class = &img2pipe_class,
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static int bmp_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
int ihsize;
|
|
|
|
|
|
|
|
if (AV_RB16(b) != 0x424d)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
ihsize = AV_RL32(b+14);
|
|
|
|
if (ihsize < 12 || ihsize > 255)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!AV_RN32(b + 6)) {
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
}
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int dds_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if ( AV_RB64(b) == 0x444453207c000000
|
|
|
|
&& AV_RL32(b + 8)
|
|
|
|
&& AV_RL32(b + 12))
|
|
|
|
return AVPROBE_SCORE_MAX - 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int dpx_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
int w, h;
|
|
|
|
int is_big = (AV_RN32(b) == AV_RN32("SDPX"));
|
|
|
|
|
|
|
|
if (p->buf_size < 0x304+8)
|
|
|
|
return 0;
|
|
|
|
w = is_big ? AV_RB32(p->buf + 0x304) : AV_RL32(p->buf + 0x304);
|
|
|
|
h = is_big ? AV_RB32(p->buf + 0x308) : AV_RL32(p->buf + 0x308);
|
|
|
|
if (w <= 0 || h <= 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (is_big || AV_RN32(b) == AV_RN32("XPDS"))
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int exr_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RL32(b) == 20000630)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int j2k_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB64(b) == 0x0000000c6a502020 ||
|
|
|
|
AV_RB32(b) == 0xff4fff51)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int jpeg_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
int i, state = SOI;
|
|
|
|
|
|
|
|
if (AV_RB16(b) != 0xFFD8 ||
|
|
|
|
AV_RB32(b) == 0xFFD8FFF7)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
b += 2;
|
|
|
|
for (i = 0; i < p->buf_size - 3; i++) {
|
|
|
|
int c;
|
|
|
|
if (b[i] != 0xFF)
|
|
|
|
continue;
|
|
|
|
c = b[i + 1];
|
|
|
|
switch (c) {
|
|
|
|
case SOI:
|
|
|
|
return 0;
|
|
|
|
case SOF0:
|
|
|
|
case SOF1:
|
|
|
|
case SOF2:
|
|
|
|
case SOF3:
|
|
|
|
case SOF5:
|
|
|
|
case SOF6:
|
|
|
|
case SOF7:
|
|
|
|
i += AV_RB16(&b[i + 2]) + 1;
|
|
|
|
if (state != SOI)
|
|
|
|
return 0;
|
|
|
|
state = SOF0;
|
|
|
|
break;
|
|
|
|
case SOS:
|
|
|
|
i += AV_RB16(&b[i + 2]) + 1;
|
|
|
|
if (state != SOF0 && state != SOS)
|
|
|
|
return 0;
|
|
|
|
state = SOS;
|
|
|
|
break;
|
|
|
|
case EOI:
|
|
|
|
if (state != SOS)
|
|
|
|
return 0;
|
|
|
|
state = EOI;
|
|
|
|
break;
|
|
|
|
case DQT:
|
|
|
|
case APP0:
|
|
|
|
case APP1:
|
|
|
|
case APP2:
|
|
|
|
case APP3:
|
|
|
|
case APP4:
|
|
|
|
case APP5:
|
|
|
|
case APP6:
|
|
|
|
case APP7:
|
|
|
|
case APP8:
|
|
|
|
case APP9:
|
|
|
|
case APP10:
|
|
|
|
case APP11:
|
|
|
|
case APP12:
|
|
|
|
case APP13:
|
|
|
|
case APP14:
|
|
|
|
case APP15:
|
|
|
|
case COM:
|
|
|
|
i += AV_RB16(&b[i + 2]) + 1;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
if ( (c > TEM && c < SOF0)
|
|
|
|
|| c == JPG)
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (state == EOI)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
if (state == SOS)
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 2;
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int jpegls_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB32(b) == 0xffd8fff7)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pcx_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if ( p->buf_size < 128
|
|
|
|
|| b[0] != 10
|
|
|
|
|| b[1] > 5
|
|
|
|
|| b[2] > 1
|
|
|
|
|| av_popcount(b[3]) != 1 || b[3] > 8
|
|
|
|
|| AV_RL16(&b[4]) > AV_RL16(&b[8])
|
|
|
|
|| AV_RL16(&b[6]) > AV_RL16(&b[10])
|
|
|
|
|| b[64])
|
|
|
|
return 0;
|
|
|
|
b += 73;
|
|
|
|
while (++b < p->buf + 128)
|
|
|
|
if (*b)
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 4;
|
|
|
|
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int qdraw_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if ( p->buf_size >= 528
|
|
|
|
&& (AV_RB64(b + 520) & 0xFFFFFFFFFFFF) == 0x001102ff0c00
|
|
|
|
&& AV_RB16(b + 520)
|
|
|
|
&& AV_RB16(b + 518))
|
|
|
|
return AVPROBE_SCORE_MAX * 3 / 4;
|
|
|
|
if ( (AV_RB64(b + 8) & 0xFFFFFFFFFFFF) == 0x001102ff0c00
|
|
|
|
&& AV_RB16(b + 8)
|
|
|
|
&& AV_RB16(b + 6))
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 4;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pictor_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RL16(b) == 0x1234)
|
|
|
|
return AVPROBE_SCORE_EXTENSION / 4;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int png_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB64(b) == 0x89504e470d0a1a0a)
|
|
|
|
return AVPROBE_SCORE_MAX - 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int psd_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
int ret = 0;
|
|
|
|
uint16_t color_mode;
|
|
|
|
|
|
|
|
if (AV_RL32(b) == MKTAG('8','B','P','S')) {
|
|
|
|
ret += 1;
|
|
|
|
} else {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((b[4] == 0) && (b[5] == 1)) {/* version 1 is PSD, version 2 is PSB */
|
|
|
|
ret += 1;
|
|
|
|
} else {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((AV_RL32(b+6) == 0) && (AV_RL16(b+10) == 0))/* reserved must be 0 */
|
|
|
|
ret += 1;
|
|
|
|
|
|
|
|
color_mode = AV_RB16(b+24);
|
|
|
|
if ((color_mode <= 9) && (color_mode != 5) && (color_mode != 6))
|
|
|
|
ret += 1;
|
|
|
|
|
|
|
|
return AVPROBE_SCORE_EXTENSION + ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int sgi_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB16(b) == 474 &&
|
|
|
|
(b[2] & ~1) == 0 &&
|
|
|
|
(b[3] & ~3) == 0 && b[3] &&
|
|
|
|
(AV_RB16(b + 4) & ~7) == 0 && AV_RB16(b + 4))
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int sunrast_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB32(b) == 0x59a66a95)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int svg_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
const uint8_t *end = p->buf + p->buf_size;
|
|
|
|
|
|
|
|
if (memcmp(p->buf, "<?xml", 5))
|
|
|
|
return 0;
|
|
|
|
while (b < end) {
|
|
|
|
int inc = ff_subtitles_next_line(b);
|
|
|
|
if (!inc)
|
|
|
|
break;
|
|
|
|
b += inc;
|
|
|
|
if (b >= end - 4)
|
|
|
|
return 0;
|
|
|
|
if (!memcmp(b, "<svg", 4))
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tiff_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB32(b) == 0x49492a00 ||
|
|
|
|
AV_RB32(b) == 0x4D4D002a)
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int webp_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB32(b) == 0x52494646 &&
|
|
|
|
AV_RB32(b + 8) == 0x57454250)
|
|
|
|
return AVPROBE_SCORE_MAX - 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pnm_magic_check(const AVProbeData *p, int magic)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
return b[0] == 'P' && b[1] == magic + '0';
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pnm_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
while (b[2] == '\r')
|
|
|
|
b++;
|
|
|
|
if (b[2] == '\n' && (b[3] == '#' || (b[3] >= '0' && b[3] <= '9')))
|
|
|
|
return AVPROBE_SCORE_EXTENSION + 2;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pbm_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
return pnm_magic_check(p, 1) || pnm_magic_check(p, 4) ? pnm_probe(p) : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pgmx_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
return pnm_magic_check(p, 2) || pnm_magic_check(p, 5) ? pnm_probe(p) : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pgm_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
int ret = pgmx_probe(p);
|
|
|
|
return ret && !av_match_ext(p->filename, "pgmyuv") ? ret : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pgmyuv_probe(const AVProbeData *p) // custom FFmpeg format recognized by file extension
|
|
|
|
{
|
|
|
|
int ret = pgmx_probe(p);
|
|
|
|
return ret && av_match_ext(p->filename, "pgmyuv") ? ret : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ppm_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
return pnm_magic_check(p, 3) || pnm_magic_check(p, 6) ? pnm_probe(p) : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int pam_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
return pnm_magic_check(p, 7) ? pnm_probe(p) : 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int xpm_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
|
|
|
|
if (AV_RB64(b) == 0x2f2a2058504d202a && *(b+8) == '/')
|
|
|
|
return AVPROBE_SCORE_MAX - 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int xwd_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
const uint8_t *b = p->buf;
|
|
|
|
unsigned width, bpp, bpad, lsize;
|
|
|
|
|
|
|
|
if ( p->buf_size < XWD_HEADER_SIZE
|
|
|
|
|| AV_RB32(b ) < XWD_HEADER_SIZE // header size
|
|
|
|
|| AV_RB32(b + 4) != XWD_VERSION // version
|
|
|
|
|| AV_RB32(b + 8) != XWD_Z_PIXMAP // format
|
|
|
|
|| AV_RB32(b + 12) > 32 || !AV_RB32(b + 12) // depth
|
|
|
|
|| AV_RB32(b + 16) == 0 // width
|
|
|
|
|| AV_RB32(b + 20) == 0 // height
|
|
|
|
|| AV_RB32(b + 28) > 1 // byteorder
|
|
|
|
|| AV_RB32(b + 32) & ~56 || av_popcount(AV_RB32(b + 32)) != 1 // bitmap unit
|
|
|
|
|| AV_RB32(b + 36) > 1 // bitorder
|
|
|
|
|| AV_RB32(b + 40) & ~56 || av_popcount(AV_RB32(b + 40)) != 1 // padding
|
|
|
|
|| AV_RB32(b + 44) > 32 || !AV_RB32(b + 44) // bpp
|
|
|
|
|| AV_RB32(b + 68) > 256) // colours
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
width = AV_RB32(b + 16);
|
|
|
|
bpad = AV_RB32(b + 40);
|
|
|
|
bpp = AV_RB32(b + 44);
|
|
|
|
lsize = AV_RB32(b + 48);
|
|
|
|
if (lsize < FFALIGN(width * bpp, bpad) >> 3)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return AVPROBE_SCORE_MAX / 2 + 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gif_probe(const AVProbeData *p)
|
|
|
|
{
|
|
|
|
/* check magick */
|
|
|
|
if (memcmp(p->buf, gif87a_sig, 6) && memcmp(p->buf, gif89a_sig, 6))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* width or height contains zero? */
|
|
|
|
if (!AV_RL16(&p->buf[6]) || !AV_RL16(&p->buf[8]))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return AVPROBE_SCORE_MAX - 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define IMAGEAUTO_DEMUXER(imgname, codecid)\
|
|
|
|
static const AVClass imgname ## _class = {\
|
|
|
|
.class_name = AV_STRINGIFY(imgname) " demuxer",\
|
|
|
|
.item_name = av_default_item_name,\
|
|
|
|
.option = ff_img2pipe_options,\
|
|
|
|
.version = LIBAVUTIL_VERSION_INT,\
|
|
|
|
};\
|
|
|
|
AVInputFormat ff_image_ ## imgname ## _pipe_demuxer = {\
|
|
|
|
.name = AV_STRINGIFY(imgname) "_pipe",\
|
|
|
|
.long_name = NULL_IF_CONFIG_SMALL("piped " AV_STRINGIFY(imgname) " sequence"),\
|
|
|
|
.priv_data_size = sizeof(VideoDemuxData),\
|
|
|
|
.read_probe = imgname ## _probe,\
|
|
|
|
.read_header = ff_img_read_header,\
|
|
|
|
.read_packet = ff_img_read_packet,\
|
|
|
|
.priv_class = & imgname ## _class,\
|
|
|
|
.flags = AVFMT_GENERIC_INDEX, \
|
|
|
|
.raw_codec_id = codecid,\
|
|
|
|
};
|
|
|
|
|
|
|
|
IMAGEAUTO_DEMUXER(bmp, AV_CODEC_ID_BMP)
|
|
|
|
IMAGEAUTO_DEMUXER(dds, AV_CODEC_ID_DDS)
|
|
|
|
IMAGEAUTO_DEMUXER(dpx, AV_CODEC_ID_DPX)
|
|
|
|
IMAGEAUTO_DEMUXER(exr, AV_CODEC_ID_EXR)
|
|
|
|
IMAGEAUTO_DEMUXER(gif, AV_CODEC_ID_GIF)
|
|
|
|
IMAGEAUTO_DEMUXER(j2k, AV_CODEC_ID_JPEG2000)
|
|
|
|
IMAGEAUTO_DEMUXER(jpeg, AV_CODEC_ID_MJPEG)
|
|
|
|
IMAGEAUTO_DEMUXER(jpegls, AV_CODEC_ID_JPEGLS)
|
|
|
|
IMAGEAUTO_DEMUXER(pam, AV_CODEC_ID_PAM)
|
|
|
|
IMAGEAUTO_DEMUXER(pbm, AV_CODEC_ID_PBM)
|
|
|
|
IMAGEAUTO_DEMUXER(pcx, AV_CODEC_ID_PCX)
|
|
|
|
IMAGEAUTO_DEMUXER(pgm, AV_CODEC_ID_PGM)
|
|
|
|
IMAGEAUTO_DEMUXER(pgmyuv, AV_CODEC_ID_PGMYUV)
|
|
|
|
IMAGEAUTO_DEMUXER(pictor, AV_CODEC_ID_PICTOR)
|
|
|
|
IMAGEAUTO_DEMUXER(png, AV_CODEC_ID_PNG)
|
|
|
|
IMAGEAUTO_DEMUXER(ppm, AV_CODEC_ID_PPM)
|
|
|
|
IMAGEAUTO_DEMUXER(psd, AV_CODEC_ID_PSD)
|
|
|
|
IMAGEAUTO_DEMUXER(qdraw, AV_CODEC_ID_QDRAW)
|
|
|
|
IMAGEAUTO_DEMUXER(sgi, AV_CODEC_ID_SGI)
|
|
|
|
IMAGEAUTO_DEMUXER(sunrast, AV_CODEC_ID_SUNRAST)
|
|
|
|
IMAGEAUTO_DEMUXER(svg, AV_CODEC_ID_SVG)
|
|
|
|
IMAGEAUTO_DEMUXER(tiff, AV_CODEC_ID_TIFF)
|
|
|
|
IMAGEAUTO_DEMUXER(webp, AV_CODEC_ID_WEBP)
|
|
|
|
IMAGEAUTO_DEMUXER(xpm, AV_CODEC_ID_XPM)
|
|
|
|
IMAGEAUTO_DEMUXER(xwd, AV_CODEC_ID_XWD)
|