|
|
|
/*
|
|
|
|
* VFW capture interface
|
|
|
|
* Copyright (c) 2006-2008 Ramiro Polla
|
|
|
|
*
|
|
|
|
* This file is part of FFmpeg.
|
|
|
|
*
|
|
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "libavutil/internal.h"
|
|
|
|
#include "libavutil/log.h"
|
|
|
|
#include "libavutil/opt.h"
|
|
|
|
#include "libavutil/parseutils.h"
|
|
|
|
|
|
|
|
#include "libavcodec/packet_internal.h"
|
|
|
|
#include "libavformat/demux.h"
|
|
|
|
#include "libavformat/internal.h"
|
|
|
|
|
|
|
|
// windows.h must no be included before winsock2.h, and libavformat internal
|
|
|
|
// headers may include winsock2.h
|
|
|
|
#include <windows.h>
|
|
|
|
// windows.h needs to be included before vfw.h
|
|
|
|
#include <vfw.h>
|
|
|
|
|
|
|
|
#include "avdevice.h"
|
|
|
|
|
|
|
|
/* Some obsolete versions of MinGW32 before 4.0.0 lack this. */
|
|
|
|
#ifndef HWND_MESSAGE
|
|
|
|
#define HWND_MESSAGE ((HWND) -3)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
struct vfw_ctx {
|
|
|
|
const AVClass *class;
|
|
|
|
HWND hwnd;
|
|
|
|
HANDLE mutex;
|
|
|
|
HANDLE event;
|
|
|
|
PacketListEntry *pktl;
|
|
|
|
unsigned int curbufsize;
|
|
|
|
unsigned int frame_num;
|
|
|
|
char *video_size; /**< A string describing video size, set by a private option. */
|
|
|
|
char *framerate; /**< Set by a private option. */
|
|
|
|
};
|
|
|
|
|
|
|
|
static enum AVPixelFormat vfw_pixfmt(DWORD biCompression, WORD biBitCount)
|
|
|
|
{
|
|
|
|
switch(biCompression) {
|
|
|
|
case MKTAG('U', 'Y', 'V', 'Y'):
|
|
|
|
return AV_PIX_FMT_UYVY422;
|
|
|
|
case MKTAG('Y', 'U', 'Y', '2'):
|
|
|
|
return AV_PIX_FMT_YUYV422;
|
|
|
|
case MKTAG('I', '4', '2', '0'):
|
|
|
|
return AV_PIX_FMT_YUV420P;
|
|
|
|
case BI_RGB:
|
|
|
|
switch(biBitCount) { /* 1-8 are untested */
|
|
|
|
case 1:
|
|
|
|
return AV_PIX_FMT_MONOWHITE;
|
|
|
|
case 4:
|
|
|
|
return AV_PIX_FMT_RGB4;
|
|
|
|
case 8:
|
|
|
|
return AV_PIX_FMT_RGB8;
|
|
|
|
case 16:
|
|
|
|
return AV_PIX_FMT_RGB555;
|
|
|
|
case 24:
|
|
|
|
return AV_PIX_FMT_BGR24;
|
|
|
|
case 32:
|
|
|
|
return AV_PIX_FMT_RGB32;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return AV_PIX_FMT_NONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static enum AVCodecID vfw_codecid(DWORD biCompression)
|
|
|
|
{
|
|
|
|
switch(biCompression) {
|
|
|
|
case MKTAG('d', 'v', 's', 'd'):
|
|
|
|
return AV_CODEC_ID_DVVIDEO;
|
|
|
|
case MKTAG('M', 'J', 'P', 'G'):
|
|
|
|
case MKTAG('m', 'j', 'p', 'g'):
|
|
|
|
return AV_CODEC_ID_MJPEG;
|
|
|
|
}
|
|
|
|
return AV_CODEC_ID_NONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define dstruct(pctx, sname, var, type) \
|
|
|
|
av_log(pctx, AV_LOG_DEBUG, #var":\t%"type"\n", sname->var)
|
|
|
|
|
|
|
|
static void dump_captureparms(AVFormatContext *s, CAPTUREPARMS *cparms)
|
|
|
|
{
|
|
|
|
av_log(s, AV_LOG_DEBUG, "CAPTUREPARMS\n");
|
|
|
|
dstruct(s, cparms, dwRequestMicroSecPerFrame, "lu");
|
|
|
|
dstruct(s, cparms, fMakeUserHitOKToCapture, "d");
|
|
|
|
dstruct(s, cparms, wPercentDropForError, "u");
|
|
|
|
dstruct(s, cparms, fYield, "d");
|
|
|
|
dstruct(s, cparms, dwIndexSize, "lu");
|
|
|
|
dstruct(s, cparms, wChunkGranularity, "u");
|
|
|
|
dstruct(s, cparms, fUsingDOSMemory, "d");
|
|
|
|
dstruct(s, cparms, wNumVideoRequested, "u");
|
|
|
|
dstruct(s, cparms, fCaptureAudio, "d");
|
|
|
|
dstruct(s, cparms, wNumAudioRequested, "u");
|
|
|
|
dstruct(s, cparms, vKeyAbort, "u");
|
|
|
|
dstruct(s, cparms, fAbortLeftMouse, "d");
|
|
|
|
dstruct(s, cparms, fAbortRightMouse, "d");
|
|
|
|
dstruct(s, cparms, fLimitEnabled, "d");
|
|
|
|
dstruct(s, cparms, wTimeLimit, "u");
|
|
|
|
dstruct(s, cparms, fMCIControl, "d");
|
|
|
|
dstruct(s, cparms, fStepMCIDevice, "d");
|
|
|
|
dstruct(s, cparms, dwMCIStartTime, "lu");
|
|
|
|
dstruct(s, cparms, dwMCIStopTime, "lu");
|
|
|
|
dstruct(s, cparms, fStepCaptureAt2x, "d");
|
|
|
|
dstruct(s, cparms, wStepCaptureAverageFrames, "u");
|
|
|
|
dstruct(s, cparms, dwAudioBufferSize, "lu");
|
|
|
|
dstruct(s, cparms, fDisableWriteCache, "d");
|
|
|
|
dstruct(s, cparms, AVStreamMaster, "u");
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dump_videohdr(AVFormatContext *s, VIDEOHDR *vhdr)
|
|
|
|
{
|
|
|
|
#ifdef DEBUG
|
|
|
|
av_log(s, AV_LOG_DEBUG, "VIDEOHDR\n");
|
|
|
|
dstruct(s, vhdr, lpData, "p");
|
|
|
|
dstruct(s, vhdr, dwBufferLength, "lu");
|
|
|
|
dstruct(s, vhdr, dwBytesUsed, "lu");
|
|
|
|
dstruct(s, vhdr, dwTimeCaptured, "lu");
|
|
|
|
dstruct(s, vhdr, dwUser, "lu");
|
|
|
|
dstruct(s, vhdr, dwFlags, "lu");
|
|
|
|
dstruct(s, vhdr, dwReserved[0], "lu");
|
|
|
|
dstruct(s, vhdr, dwReserved[1], "lu");
|
|
|
|
dstruct(s, vhdr, dwReserved[2], "lu");
|
|
|
|
dstruct(s, vhdr, dwReserved[3], "lu");
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dump_bih(AVFormatContext *s, BITMAPINFOHEADER *bih)
|
|
|
|
{
|
|
|
|
av_log(s, AV_LOG_DEBUG, "BITMAPINFOHEADER\n");
|
|
|
|
dstruct(s, bih, biSize, "lu");
|
|
|
|
dstruct(s, bih, biWidth, "ld");
|
|
|
|
dstruct(s, bih, biHeight, "ld");
|
|
|
|
dstruct(s, bih, biPlanes, "d");
|
|
|
|
dstruct(s, bih, biBitCount, "d");
|
|
|
|
dstruct(s, bih, biCompression, "lu");
|
|
|
|
av_log(s, AV_LOG_DEBUG, " biCompression:\t\"%.4s\"\n",
|
|
|
|
(char*) &bih->biCompression);
|
|
|
|
dstruct(s, bih, biSizeImage, "lu");
|
|
|
|
dstruct(s, bih, biXPelsPerMeter, "lu");
|
|
|
|
dstruct(s, bih, biYPelsPerMeter, "lu");
|
|
|
|
dstruct(s, bih, biClrUsed, "lu");
|
|
|
|
dstruct(s, bih, biClrImportant, "lu");
|
|
|
|
}
|
|
|
|
|
|
|
|
static int shall_we_drop(AVFormatContext *s)
|
|
|
|
{
|
|
|
|
struct vfw_ctx *ctx = s->priv_data;
|
|
|
|
static const uint8_t dropscore[4] = { 62, 75, 87, 100 };
|
|
|
|
const int ndropscores = FF_ARRAY_ELEMS(dropscore);
|
|
|
|
unsigned int buffer_fullness = (ctx->curbufsize*100)/s->max_picture_buffer;
|
|
|
|
|
|
|
|
if(dropscore[++ctx->frame_num%ndropscores] <= buffer_fullness) {
|
|
|
|
av_log(s, AV_LOG_ERROR,
|
|
|
|
"real-time buffer %d%% full! frame dropped!\n", buffer_fullness);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static LRESULT CALLBACK videostream_cb(HWND hwnd, LPVIDEOHDR vdhdr)
|
|
|
|
{
|
|
|
|
AVFormatContext *s;
|
|
|
|
struct vfw_ctx *ctx;
|
|
|
|
PacketListEntry **ppktl, *pktl_next;
|
|
|
|
|
|
|
|
s = (AVFormatContext *) GetWindowLongPtr(hwnd, GWLP_USERDATA);
|
|
|
|
ctx = s->priv_data;
|
|
|
|
|
|
|
|
dump_videohdr(s, vdhdr);
|
|
|
|
|
|
|
|
if(shall_we_drop(s))
|
|
|
|
return FALSE;
|
|
|
|
|
|
|
|
WaitForSingleObject(ctx->mutex, INFINITE);
|
|
|
|
|
|
|
|
pktl_next = av_mallocz(sizeof(*pktl_next));
|
|
|
|
if(!pktl_next)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
if(av_new_packet(&pktl_next->pkt, vdhdr->dwBytesUsed) < 0) {
|
|
|
|
av_free(pktl_next);
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
pktl_next->pkt.pts = vdhdr->dwTimeCaptured;
|
|
|
|
memcpy(pktl_next->pkt.data, vdhdr->lpData, vdhdr->dwBytesUsed);
|
|
|
|
|
|
|
|
for(ppktl = &ctx->pktl ; *ppktl ; ppktl = &(*ppktl)->next);
|
|
|
|
*ppktl = pktl_next;
|
|
|
|
|
|
|
|
ctx->curbufsize += vdhdr->dwBytesUsed;
|
|
|
|
|
|
|
|
SetEvent(ctx->event);
|
|
|
|
ReleaseMutex(ctx->mutex);
|
|
|
|
|
|
|
|
return TRUE;
|
|
|
|
fail:
|
|
|
|
ReleaseMutex(ctx->mutex);
|
|
|
|
return FALSE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int vfw_read_close(AVFormatContext *s)
|
|
|
|
{
|
|
|
|
struct vfw_ctx *ctx = s->priv_data;
|
|
|
|
PacketListEntry *pktl;
|
|
|
|
|
|
|
|
if(ctx->hwnd) {
|
|
|
|
SendMessage(ctx->hwnd, WM_CAP_SET_CALLBACK_VIDEOSTREAM, 0, 0);
|
|
|
|
SendMessage(ctx->hwnd, WM_CAP_DRIVER_DISCONNECT, 0, 0);
|
|
|
|
DestroyWindow(ctx->hwnd);
|
|
|
|
}
|
|
|
|
if(ctx->mutex)
|
|
|
|
CloseHandle(ctx->mutex);
|
|
|
|
if(ctx->event)
|
|
|
|
CloseHandle(ctx->event);
|
|
|
|
|
|
|
|
pktl = ctx->pktl;
|
|
|
|
while (pktl) {
|
|
|
|
PacketListEntry *next = pktl->next;
|
|
|
|
av_packet_unref(&pktl->pkt);
|
|
|
|
av_free(pktl);
|
|
|
|
pktl = next;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int vfw_read_header(AVFormatContext *s)
|
|
|
|
{
|
|
|
|
struct vfw_ctx *ctx = s->priv_data;
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
AVCodecParameters *par;
|
|
|
|
AVStream *st;
|
|
|
|
int devnum;
|
|
|
|
int bisize;
|
|
|
|
BITMAPINFO *bi = NULL;
|
|
|
|
CAPTUREPARMS cparms;
|
|
|
|
DWORD biCompression;
|
|
|
|
WORD biBitCount;
|
|
|
|
int ret;
|
|
|
|
AVRational framerate_q;
|
|
|
|
|
|
|
|
if (!strcmp(s->url, "list")) {
|
|
|
|
for (devnum = 0; devnum <= 9; devnum++) {
|
|
|
|
char driver_name[256];
|
|
|
|
char driver_ver[256];
|
|
|
|
ret = capGetDriverDescription(devnum,
|
|
|
|
driver_name, sizeof(driver_name),
|
|
|
|
driver_ver, sizeof(driver_ver));
|
|
|
|
if (ret) {
|
|
|
|
av_log(s, AV_LOG_INFO, "Driver %d\n", devnum);
|
|
|
|
av_log(s, AV_LOG_INFO, " %s\n", driver_name);
|
|
|
|
av_log(s, AV_LOG_INFO, " %s\n", driver_ver);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return AVERROR(EIO);
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx->hwnd = capCreateCaptureWindow(NULL, 0, 0, 0, 0, 0, HWND_MESSAGE, 0);
|
|
|
|
if(!ctx->hwnd) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not create capture window.\n");
|
|
|
|
return AVERROR(EIO);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If atoi fails, devnum==0 and the default device is used */
|
|
|
|
devnum = atoi(s->url);
|
|
|
|
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_DRIVER_CONNECT, devnum, 0);
|
|
|
|
if(!ret) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not connect to device.\n");
|
|
|
|
DestroyWindow(ctx->hwnd);
|
|
|
|
return AVERROR(ENODEV);
|
|
|
|
}
|
|
|
|
|
|
|
|
SendMessage(ctx->hwnd, WM_CAP_SET_OVERLAY, 0, 0);
|
|
|
|
SendMessage(ctx->hwnd, WM_CAP_SET_PREVIEW, 0, 0);
|
|
|
|
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_SET_CALLBACK_VIDEOSTREAM, 0,
|
|
|
|
(LPARAM) videostream_cb);
|
|
|
|
if(!ret) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not set video stream callback.\n");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
SetWindowLongPtr(ctx->hwnd, GWLP_USERDATA, (LONG_PTR) s);
|
|
|
|
|
|
|
|
st = avformat_new_stream(s, NULL);
|
|
|
|
if(!st) {
|
|
|
|
vfw_read_close(s);
|
|
|
|
return AVERROR(ENOMEM);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Set video format */
|
|
|
|
bisize = SendMessage(ctx->hwnd, WM_CAP_GET_VIDEOFORMAT, 0, 0);
|
|
|
|
if(!bisize)
|
|
|
|
goto fail;
|
|
|
|
bi = av_malloc(bisize);
|
|
|
|
if(!bi) {
|
|
|
|
vfw_read_close(s);
|
|
|
|
return AVERROR(ENOMEM);
|
|
|
|
}
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_GET_VIDEOFORMAT, bisize, (LPARAM) bi);
|
|
|
|
if(!ret)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
dump_bih(s, &bi->bmiHeader);
|
|
|
|
|
|
|
|
ret = av_parse_video_rate(&framerate_q, ctx->framerate);
|
|
|
|
if (ret < 0) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not parse framerate '%s'.\n", ctx->framerate);
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ctx->video_size) {
|
|
|
|
int w, h;
|
|
|
|
ret = av_parse_video_size(&w, &h, ctx->video_size);
|
|
|
|
if (ret < 0) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Couldn't parse video size.\n");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
bi->bmiHeader.biWidth = w;
|
|
|
|
bi->bmiHeader.biHeight = h;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (0) {
|
|
|
|
/* For testing yet unsupported compressions
|
|
|
|
* Copy these values from user-supplied verbose information */
|
|
|
|
bi->bmiHeader.biWidth = 320;
|
|
|
|
bi->bmiHeader.biHeight = 240;
|
|
|
|
bi->bmiHeader.biPlanes = 1;
|
|
|
|
bi->bmiHeader.biBitCount = 12;
|
|
|
|
bi->bmiHeader.biCompression = MKTAG('I','4','2','0');
|
|
|
|
bi->bmiHeader.biSizeImage = 115200;
|
|
|
|
dump_bih(s, &bi->bmiHeader);
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_SET_VIDEOFORMAT, bisize, (LPARAM) bi);
|
|
|
|
if(!ret) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not set Video Format.\n");
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
biCompression = bi->bmiHeader.biCompression;
|
|
|
|
biBitCount = bi->bmiHeader.biBitCount;
|
|
|
|
|
|
|
|
/* Set sequence setup */
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_GET_SEQUENCE_SETUP, sizeof(cparms),
|
|
|
|
(LPARAM) &cparms);
|
|
|
|
if(!ret)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
dump_captureparms(s, &cparms);
|
|
|
|
|
|
|
|
cparms.fYield = 1; // Spawn a background thread
|
|
|
|
cparms.dwRequestMicroSecPerFrame =
|
|
|
|
(framerate_q.den*1000000) / framerate_q.num;
|
|
|
|
cparms.fAbortLeftMouse = 0;
|
|
|
|
cparms.fAbortRightMouse = 0;
|
|
|
|
cparms.fCaptureAudio = 0;
|
|
|
|
cparms.vKeyAbort = 0;
|
|
|
|
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_SET_SEQUENCE_SETUP, sizeof(cparms),
|
|
|
|
(LPARAM) &cparms);
|
|
|
|
if(!ret)
|
|
|
|
goto fail;
|
|
|
|
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
st->avg_frame_rate = framerate_q;
|
|
|
|
|
|
|
|
par = st->codecpar;
|
|
|
|
par->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
|
|
par->width = bi->bmiHeader.biWidth;
|
|
|
|
par->height = bi->bmiHeader.biHeight;
|
|
|
|
par->format = vfw_pixfmt(biCompression, biBitCount);
|
|
|
|
if (par->format == AV_PIX_FMT_NONE) {
|
|
|
|
par->codec_id = vfw_codecid(biCompression);
|
|
|
|
if (par->codec_id == AV_CODEC_ID_NONE) {
|
|
|
|
avpriv_report_missing_feature(s, "This compression type");
|
|
|
|
vfw_read_close(s);
|
|
|
|
return AVERROR_PATCHWELCOME;
|
|
|
|
}
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
par->bits_per_coded_sample = biBitCount;
|
|
|
|
} else {
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
par->codec_id = AV_CODEC_ID_RAWVIDEO;
|
|
|
|
if(biCompression == BI_RGB) {
|
lavf: replace AVStream.codec with AVStream.codecpar
Currently, AVStream contains an embedded AVCodecContext instance, which
is used by demuxers to export stream parameters to the caller and by
muxers to receive stream parameters from the caller. It is also used
internally as the codec context that is passed to parsers.
In addition, it is also widely used by the callers as the decoding (when
demuxer) or encoding (when muxing) context, though this has been
officially discouraged since Libav 11.
There are multiple important problems with this approach:
- the fields in AVCodecContext are in general one of
* stream parameters
* codec options
* codec state
However, it's not clear which ones are which. It is consequently
unclear which fields are a demuxer allowed to set or a muxer allowed to
read. This leads to erratic behaviour depending on whether decoding or
encoding is being performed or not (and whether it uses the AVStream
embedded codec context).
- various synchronization issues arising from the fact that the same
context is used by several different APIs (muxers/demuxers,
parsers, bitstream filters and encoders/decoders) simultaneously, with
there being no clear rules for who can modify what and the different
processes being typically delayed with respect to each other.
- avformat_find_stream_info() making it necessary to support opening
and closing a single codec context multiple times, thus
complicating the semantics of freeing various allocated objects in the
codec context.
Those problems are resolved by replacing the AVStream embedded codec
context with a newly added AVCodecParameters instance, which stores only
the stream parameters exported by the demuxers or read by the muxers.
11 years ago
|
|
|
par->bits_per_coded_sample = biBitCount;
|
|
|
|
par->extradata = av_malloc(9 + AV_INPUT_BUFFER_PADDING_SIZE);
|
|
|
|
if (par->extradata) {
|
|
|
|
par->extradata_size = 9;
|
|
|
|
memcpy(par->extradata, "BottomUp", 9);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
av_freep(&bi);
|
|
|
|
|
|
|
|
avpriv_set_pts_info(st, 32, 1, 1000);
|
|
|
|
|
|
|
|
ctx->mutex = CreateMutex(NULL, 0, NULL);
|
|
|
|
if(!ctx->mutex) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not create Mutex.\n" );
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
ctx->event = CreateEvent(NULL, 1, 0, NULL);
|
|
|
|
if(!ctx->event) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not create Event.\n" );
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = SendMessage(ctx->hwnd, WM_CAP_SEQUENCE_NOFILE, 0, 0);
|
|
|
|
if(!ret) {
|
|
|
|
av_log(s, AV_LOG_ERROR, "Could not start capture sequence.\n" );
|
|
|
|
goto fail;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
fail:
|
|
|
|
av_freep(&bi);
|
|
|
|
vfw_read_close(s);
|
|
|
|
return AVERROR(EIO);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int vfw_read_packet(AVFormatContext *s, AVPacket *pkt)
|
|
|
|
{
|
|
|
|
struct vfw_ctx *ctx = s->priv_data;
|
|
|
|
PacketListEntry *pktl = NULL;
|
|
|
|
|
|
|
|
while(!pktl) {
|
|
|
|
WaitForSingleObject(ctx->mutex, INFINITE);
|
|
|
|
pktl = ctx->pktl;
|
|
|
|
if(ctx->pktl) {
|
|
|
|
*pkt = ctx->pktl->pkt;
|
|
|
|
ctx->pktl = ctx->pktl->next;
|
|
|
|
av_free(pktl);
|
|
|
|
}
|
|
|
|
ResetEvent(ctx->event);
|
|
|
|
ReleaseMutex(ctx->mutex);
|
|
|
|
if(!pktl) {
|
|
|
|
if(s->flags & AVFMT_FLAG_NONBLOCK) {
|
|
|
|
return AVERROR(EAGAIN);
|
|
|
|
} else {
|
|
|
|
WaitForSingleObject(ctx->event, INFINITE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx->curbufsize -= pkt->size;
|
|
|
|
|
|
|
|
return pkt->size;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define OFFSET(x) offsetof(struct vfw_ctx, x)
|
|
|
|
#define DEC AV_OPT_FLAG_DECODING_PARAM
|
|
|
|
static const AVOption options[] = {
|
|
|
|
{ "video_size", "A string describing frame size, such as 640x480 or hd720.", OFFSET(video_size), AV_OPT_TYPE_STRING, {.str = NULL}, 0, 0, DEC },
|
|
|
|
{ "framerate", "", OFFSET(framerate), AV_OPT_TYPE_STRING, {.str = "ntsc"}, 0, 0, DEC },
|
|
|
|
{ NULL },
|
|
|
|
};
|
|
|
|
|
|
|
|
static const AVClass vfw_class = {
|
|
|
|
.class_name = "VFW indev",
|
|
|
|
.item_name = av_default_item_name,
|
|
|
|
.option = options,
|
|
|
|
.version = LIBAVUTIL_VERSION_INT,
|
|
|
|
.category = AV_CLASS_CATEGORY_DEVICE_VIDEO_INPUT
|
|
|
|
};
|
|
|
|
|
|
|
|
const FFInputFormat ff_vfwcap_demuxer = {
|
|
|
|
.p.name = "vfwcap",
|
|
|
|
.p.long_name = NULL_IF_CONFIG_SMALL("VfW video capture"),
|
|
|
|
.p.flags = AVFMT_NOFILE,
|
|
|
|
.p.priv_class = &vfw_class,
|
|
|
|
.priv_data_size = sizeof(struct vfw_ctx),
|
|
|
|
.read_header = vfw_read_header,
|
|
|
|
.read_packet = vfw_read_packet,
|
|
|
|
.read_close = vfw_read_close,
|
|
|
|
};
|