/*M/////////////////////////////////////////////////////////////////////////////////////// // // IMPORTANT: READ BEFORE DOWNLOADING, COPYING, INSTALLING OR USING. // // By downloading, copying, installing or using the software you agree to this license. // If you do not agree to this license, do not download, install, // copy or use the software. // // // Intel License Agreement // For Open Source Computer Vision Library // // Copyright (C) 2000, Intel Corporation, all rights reserved. // Third party copyrights are property of their respective owners. // // Redistribution and use in source and binary forms, with or without modification, // are permitted provided that the following conditions are met: // // * Redistribution's of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // // * Redistribution's in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // // * The name of Intel Corporation may not be used to endorse or promote products // derived from this software without specific prior written permission. // // This software is provided by the copyright holders and contributors "as is" and // any express or implied warranties, including, but not limited to, the implied // warranties of merchantability and fitness for a particular purpose are disclaimed. // In no event shall the Intel Corporation or contributors be liable for any direct, // indirect, incidental, special, exemplary, or consequential damages // (including, but not limited to, procurement of substitute goods or services; // loss of use, data, or profits; or business interruption) however caused // and on any theory of liability, whether in contract, strict liability, // or tort (including negligence or otherwise) arising in any way out of // the use of this software, even if advised of the possibility of such damage. // //M*/ #include "precomp.hpp" #if defined _WIN32 && defined HAVE_MSMF /* Media Foundation-based Video Capturing module is based on videoInput library by Evgeny Pereguda: http://www.codeproject.com/Articles/559437/Capturing-of-video-from-web-camera-on-Windows-7-an Originally licensed under The Code Project Open License (CPOL) 1.02: http://www.codeproject.com/info/cpol10.aspx */ //require Windows 8 for some of the formats defined otherwise could baseline on lower version #if WINVER < _WIN32_WINNT_WIN8 #undef WINVER #define WINVER _WIN32_WINNT_WIN8 #endif #include #include #include #include #include #include #include #include #include #ifdef HAVE_DXVA #include #include #endif #include #include #include #include #include #include #include #include #ifdef _MSC_VER #pragma warning(disable:4503) #pragma comment(lib, "mfplat") #pragma comment(lib, "mf") #pragma comment(lib, "mfuuid") #pragma comment(lib, "Strmiids") #pragma comment(lib, "Mfreadwrite") #ifdef HAVE_DXVA #pragma comment(lib, "d3d11") #endif #if (WINVER >= 0x0602) // Available since Win 8 #pragma comment(lib, "MinCore_Downlevel") #endif #endif #include #include #include // QISearch struct IMFMediaType; struct IMFActivate; struct IMFMediaSource; struct IMFAttributes; namespace { template class ComPtr { public: ComPtr() { } ComPtr(T* lp) { p = lp; } ComPtr(_In_ const ComPtr& lp) { p = lp.p; } virtual ~ComPtr() { } T** operator&() { CV_Assert(p == NULL); return p.operator&(); } T* operator->() const { CV_Assert(p != NULL); return p.operator->(); } operator bool() { return p.operator!=(NULL); } T* Get() const { return p; } void Release() { if (p) p.Release(); } // query for U interface template HRESULT As(_Out_ ComPtr& lp) const { lp.Release(); return p->QueryInterface(__uuidof(U), reinterpret_cast((T**)&lp)); } private: _COM_SMARTPTR_TYPEDEF(T, __uuidof(T)); TPtr p; }; #define _ComPtr ComPtr // Structure for collecting info about types of video, which are supported by current video device struct MediaType { unsigned int MF_MT_FRAME_SIZE; UINT32 height; UINT32 width; unsigned int MF_MT_YUV_MATRIX; unsigned int MF_MT_VIDEO_LIGHTING; int MF_MT_DEFAULT_STRIDE; // stride is negative if image is bottom-up unsigned int MF_MT_VIDEO_CHROMA_SITING; GUID MF_MT_AM_FORMAT_TYPE; unsigned int MF_MT_FIXED_SIZE_SAMPLES; unsigned int MF_MT_VIDEO_NOMINAL_RANGE; UINT32 MF_MT_FRAME_RATE_NUMERATOR; UINT32 MF_MT_FRAME_RATE_DENOMINATOR; UINT32 MF_MT_PIXEL_ASPECT_RATIO_NUMERATOR; UINT32 MF_MT_PIXEL_ASPECT_RATIO_DENOMINATOR; unsigned int MF_MT_ALL_SAMPLES_INDEPENDENT; UINT32 MF_MT_FRAME_RATE_RANGE_MIN_NUMERATOR; UINT32 MF_MT_FRAME_RATE_RANGE_MIN_DENOMINATOR; unsigned int MF_MT_SAMPLE_SIZE; unsigned int MF_MT_VIDEO_PRIMARIES; unsigned int MF_MT_INTERLACE_MODE; UINT32 MF_MT_FRAME_RATE_RANGE_MAX_NUMERATOR; UINT32 MF_MT_FRAME_RATE_RANGE_MAX_DENOMINATOR; GUID MF_MT_MAJOR_TYPE; GUID MF_MT_SUBTYPE; LPCWSTR pMF_MT_MAJOR_TYPEName; LPCWSTR pMF_MT_SUBTYPEName; MediaType(); MediaType(IMFMediaType *pType); ~MediaType(); void Clear(); }; // Class for creating of Media Foundation context class Media_Foundation { public: ~Media_Foundation(void) { /*CV_Assert(SUCCEEDED(MFShutdown()));*/ CoUninitialize(); } static Media_Foundation& getInstance() { static Media_Foundation instance; return instance; } private: Media_Foundation(void) { CoInitialize(0); CV_Assert(SUCCEEDED(MFStartup(MF_VERSION))); } }; #ifndef IF_GUID_EQUAL_RETURN #define IF_GUID_EQUAL_RETURN(val) if(val == guid) return L#val #endif LPCWSTR GetGUIDNameConstNew(const GUID& guid) { IF_GUID_EQUAL_RETURN(MF_MT_MAJOR_TYPE); IF_GUID_EQUAL_RETURN(MF_MT_SUBTYPE); IF_GUID_EQUAL_RETURN(MF_MT_ALL_SAMPLES_INDEPENDENT); IF_GUID_EQUAL_RETURN(MF_MT_FIXED_SIZE_SAMPLES); IF_GUID_EQUAL_RETURN(MF_MT_COMPRESSED); IF_GUID_EQUAL_RETURN(MF_MT_SAMPLE_SIZE); IF_GUID_EQUAL_RETURN(MF_MT_WRAPPED_TYPE); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_NUM_CHANNELS); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_SAMPLES_PER_SECOND); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_FLOAT_SAMPLES_PER_SECOND); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_AVG_BYTES_PER_SECOND); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_BLOCK_ALIGNMENT); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_BITS_PER_SAMPLE); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_VALID_BITS_PER_SAMPLE); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_SAMPLES_PER_BLOCK); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_CHANNEL_MASK); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_FOLDDOWN_MATRIX); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_WMADRC_PEAKREF); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_WMADRC_PEAKTARGET); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_WMADRC_AVGREF); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_WMADRC_AVGTARGET); IF_GUID_EQUAL_RETURN(MF_MT_AUDIO_PREFER_WAVEFORMATEX); IF_GUID_EQUAL_RETURN(MF_MT_AAC_PAYLOAD_TYPE); IF_GUID_EQUAL_RETURN(MF_MT_AAC_AUDIO_PROFILE_LEVEL_INDICATION); IF_GUID_EQUAL_RETURN(MF_MT_FRAME_SIZE); IF_GUID_EQUAL_RETURN(MF_MT_FRAME_RATE); IF_GUID_EQUAL_RETURN(MF_MT_FRAME_RATE_RANGE_MAX); IF_GUID_EQUAL_RETURN(MF_MT_FRAME_RATE_RANGE_MIN); IF_GUID_EQUAL_RETURN(MF_MT_PIXEL_ASPECT_RATIO); IF_GUID_EQUAL_RETURN(MF_MT_DRM_FLAGS); IF_GUID_EQUAL_RETURN(MF_MT_PAD_CONTROL_FLAGS); IF_GUID_EQUAL_RETURN(MF_MT_SOURCE_CONTENT_HINT); IF_GUID_EQUAL_RETURN(MF_MT_VIDEO_CHROMA_SITING); IF_GUID_EQUAL_RETURN(MF_MT_INTERLACE_MODE); IF_GUID_EQUAL_RETURN(MF_MT_TRANSFER_FUNCTION); IF_GUID_EQUAL_RETURN(MF_MT_VIDEO_PRIMARIES); IF_GUID_EQUAL_RETURN(MF_MT_CUSTOM_VIDEO_PRIMARIES); IF_GUID_EQUAL_RETURN(MF_MT_YUV_MATRIX); IF_GUID_EQUAL_RETURN(MF_MT_VIDEO_LIGHTING); IF_GUID_EQUAL_RETURN(MF_MT_VIDEO_NOMINAL_RANGE); IF_GUID_EQUAL_RETURN(MF_MT_GEOMETRIC_APERTURE); IF_GUID_EQUAL_RETURN(MF_MT_MINIMUM_DISPLAY_APERTURE); IF_GUID_EQUAL_RETURN(MF_MT_PAN_SCAN_APERTURE); IF_GUID_EQUAL_RETURN(MF_MT_PAN_SCAN_ENABLED); IF_GUID_EQUAL_RETURN(MF_MT_AVG_BITRATE); IF_GUID_EQUAL_RETURN(MF_MT_AVG_BIT_ERROR_RATE); IF_GUID_EQUAL_RETURN(MF_MT_MAX_KEYFRAME_SPACING); IF_GUID_EQUAL_RETURN(MF_MT_DEFAULT_STRIDE); IF_GUID_EQUAL_RETURN(MF_MT_PALETTE); IF_GUID_EQUAL_RETURN(MF_MT_USER_DATA); IF_GUID_EQUAL_RETURN(MF_MT_AM_FORMAT_TYPE); IF_GUID_EQUAL_RETURN(MF_MT_MPEG_START_TIME_CODE); IF_GUID_EQUAL_RETURN(MF_MT_MPEG2_PROFILE); IF_GUID_EQUAL_RETURN(MF_MT_MPEG2_LEVEL); IF_GUID_EQUAL_RETURN(MF_MT_MPEG2_FLAGS); IF_GUID_EQUAL_RETURN(MF_MT_MPEG_SEQUENCE_HEADER); IF_GUID_EQUAL_RETURN(MF_MT_DV_AAUX_SRC_PACK_0); IF_GUID_EQUAL_RETURN(MF_MT_DV_AAUX_CTRL_PACK_0); IF_GUID_EQUAL_RETURN(MF_MT_DV_AAUX_SRC_PACK_1); IF_GUID_EQUAL_RETURN(MF_MT_DV_AAUX_CTRL_PACK_1); IF_GUID_EQUAL_RETURN(MF_MT_DV_VAUX_SRC_PACK); IF_GUID_EQUAL_RETURN(MF_MT_DV_VAUX_CTRL_PACK); IF_GUID_EQUAL_RETURN(MF_MT_ARBITRARY_HEADER); IF_GUID_EQUAL_RETURN(MF_MT_ARBITRARY_FORMAT); IF_GUID_EQUAL_RETURN(MF_MT_IMAGE_LOSS_TOLERANT); IF_GUID_EQUAL_RETURN(MF_MT_MPEG4_SAMPLE_DESCRIPTION); IF_GUID_EQUAL_RETURN(MF_MT_MPEG4_CURRENT_SAMPLE_ENTRY); IF_GUID_EQUAL_RETURN(MF_MT_ORIGINAL_4CC); IF_GUID_EQUAL_RETURN(MF_MT_ORIGINAL_WAVE_FORMAT_TAG); // Media types IF_GUID_EQUAL_RETURN(MFMediaType_Audio); IF_GUID_EQUAL_RETURN(MFMediaType_Video); IF_GUID_EQUAL_RETURN(MFMediaType_Protected); #ifdef MFMediaType_Perception IF_GUID_EQUAL_RETURN(MFMediaType_Perception); #endif IF_GUID_EQUAL_RETURN(MFMediaType_Stream); IF_GUID_EQUAL_RETURN(MFMediaType_SAMI); IF_GUID_EQUAL_RETURN(MFMediaType_Script); IF_GUID_EQUAL_RETURN(MFMediaType_Image); IF_GUID_EQUAL_RETURN(MFMediaType_HTML); IF_GUID_EQUAL_RETURN(MFMediaType_Binary); IF_GUID_EQUAL_RETURN(MFMediaType_FileTransfer); IF_GUID_EQUAL_RETURN(MFVideoFormat_AI44); // FCC('AI44') IF_GUID_EQUAL_RETURN(MFVideoFormat_ARGB32); // D3DFMT_A8R8G8B8 IF_GUID_EQUAL_RETURN(MFVideoFormat_AYUV); // FCC('AYUV') IF_GUID_EQUAL_RETURN(MFVideoFormat_DV25); // FCC('dv25') IF_GUID_EQUAL_RETURN(MFVideoFormat_DV50); // FCC('dv50') IF_GUID_EQUAL_RETURN(MFVideoFormat_DVH1); // FCC('dvh1') IF_GUID_EQUAL_RETURN(MFVideoFormat_DVC); IF_GUID_EQUAL_RETURN(MFVideoFormat_DVHD); IF_GUID_EQUAL_RETURN(MFVideoFormat_DVSD); // FCC('dvsd') IF_GUID_EQUAL_RETURN(MFVideoFormat_DVSL); // FCC('dvsl') IF_GUID_EQUAL_RETURN(MFVideoFormat_H264); // FCC('H264') IF_GUID_EQUAL_RETURN(MFVideoFormat_I420); // FCC('I420') IF_GUID_EQUAL_RETURN(MFVideoFormat_IYUV); // FCC('IYUV') IF_GUID_EQUAL_RETURN(MFVideoFormat_M4S2); // FCC('M4S2') IF_GUID_EQUAL_RETURN(MFVideoFormat_MJPG); IF_GUID_EQUAL_RETURN(MFVideoFormat_MP43); // FCC('MP43') IF_GUID_EQUAL_RETURN(MFVideoFormat_MP4S); // FCC('MP4S') IF_GUID_EQUAL_RETURN(MFVideoFormat_MP4V); // FCC('MP4V') IF_GUID_EQUAL_RETURN(MFVideoFormat_MPG1); // FCC('MPG1') IF_GUID_EQUAL_RETURN(MFVideoFormat_MSS1); // FCC('MSS1') IF_GUID_EQUAL_RETURN(MFVideoFormat_MSS2); // FCC('MSS2') IF_GUID_EQUAL_RETURN(MFVideoFormat_NV11); // FCC('NV11') IF_GUID_EQUAL_RETURN(MFVideoFormat_NV12); // FCC('NV12') IF_GUID_EQUAL_RETURN(MFVideoFormat_P010); // FCC('P010') IF_GUID_EQUAL_RETURN(MFVideoFormat_P016); // FCC('P016') IF_GUID_EQUAL_RETURN(MFVideoFormat_P210); // FCC('P210') IF_GUID_EQUAL_RETURN(MFVideoFormat_P216); // FCC('P216') IF_GUID_EQUAL_RETURN(MFVideoFormat_RGB24); // D3DFMT_R8G8B8 IF_GUID_EQUAL_RETURN(MFVideoFormat_RGB32); // D3DFMT_X8R8G8B8 IF_GUID_EQUAL_RETURN(MFVideoFormat_RGB555); // D3DFMT_X1R5G5B5 IF_GUID_EQUAL_RETURN(MFVideoFormat_RGB565); // D3DFMT_R5G6B5 IF_GUID_EQUAL_RETURN(MFVideoFormat_RGB8); IF_GUID_EQUAL_RETURN(MFVideoFormat_UYVY); // FCC('UYVY') IF_GUID_EQUAL_RETURN(MFVideoFormat_v210); // FCC('v210') IF_GUID_EQUAL_RETURN(MFVideoFormat_v410); // FCC('v410') IF_GUID_EQUAL_RETURN(MFVideoFormat_WMV1); // FCC('WMV1') IF_GUID_EQUAL_RETURN(MFVideoFormat_WMV2); // FCC('WMV2') IF_GUID_EQUAL_RETURN(MFVideoFormat_WMV3); // FCC('WMV3') IF_GUID_EQUAL_RETURN(MFVideoFormat_WVC1); // FCC('WVC1') IF_GUID_EQUAL_RETURN(MFVideoFormat_Y210); // FCC('Y210') IF_GUID_EQUAL_RETURN(MFVideoFormat_Y216); // FCC('Y216') IF_GUID_EQUAL_RETURN(MFVideoFormat_Y410); // FCC('Y410') IF_GUID_EQUAL_RETURN(MFVideoFormat_Y416); // FCC('Y416') IF_GUID_EQUAL_RETURN(MFVideoFormat_Y41P); IF_GUID_EQUAL_RETURN(MFVideoFormat_Y41T); IF_GUID_EQUAL_RETURN(MFVideoFormat_YUY2); // FCC('YUY2') IF_GUID_EQUAL_RETURN(MFVideoFormat_YV12); // FCC('YV12') IF_GUID_EQUAL_RETURN(MFVideoFormat_YVYU); #ifdef MFVideoFormat_H263 IF_GUID_EQUAL_RETURN(MFVideoFormat_H263); #endif #ifdef MFVideoFormat_H265 IF_GUID_EQUAL_RETURN(MFVideoFormat_H265); #endif #ifdef MFVideoFormat_H264_ES IF_GUID_EQUAL_RETURN(MFVideoFormat_H264_ES); #endif #ifdef MFVideoFormat_HEVC IF_GUID_EQUAL_RETURN(MFVideoFormat_HEVC); #endif #ifdef MFVideoFormat_HEVC_ES IF_GUID_EQUAL_RETURN(MFVideoFormat_HEVC_ES); #endif #ifdef MFVideoFormat_MPEG2 IF_GUID_EQUAL_RETURN(MFVideoFormat_MPEG2); #endif #ifdef MFVideoFormat_VP80 IF_GUID_EQUAL_RETURN(MFVideoFormat_VP80); #endif #ifdef MFVideoFormat_VP90 IF_GUID_EQUAL_RETURN(MFVideoFormat_VP90); #endif #ifdef MFVideoFormat_420O IF_GUID_EQUAL_RETURN(MFVideoFormat_420O); #endif #ifdef MFVideoFormat_Y42T IF_GUID_EQUAL_RETURN(MFVideoFormat_Y42T); #endif #ifdef MFVideoFormat_YVU9 IF_GUID_EQUAL_RETURN(MFVideoFormat_YVU9); #endif #ifdef MFVideoFormat_v216 IF_GUID_EQUAL_RETURN(MFVideoFormat_v216); #endif #ifdef MFVideoFormat_L8 IF_GUID_EQUAL_RETURN(MFVideoFormat_L8); #endif #ifdef MFVideoFormat_L16 IF_GUID_EQUAL_RETURN(MFVideoFormat_L16); #endif #ifdef MFVideoFormat_D16 IF_GUID_EQUAL_RETURN(MFVideoFormat_D16); #endif #ifdef D3DFMT_X8R8G8B8 IF_GUID_EQUAL_RETURN(D3DFMT_X8R8G8B8); #endif #ifdef D3DFMT_A8R8G8B8 IF_GUID_EQUAL_RETURN(D3DFMT_A8R8G8B8); #endif #ifdef D3DFMT_R8G8B8 IF_GUID_EQUAL_RETURN(D3DFMT_R8G8B8); #endif #ifdef D3DFMT_X1R5G5B5 IF_GUID_EQUAL_RETURN(D3DFMT_X1R5G5B5); #endif #ifdef D3DFMT_A4R4G4B4 IF_GUID_EQUAL_RETURN(D3DFMT_A4R4G4B4); #endif #ifdef D3DFMT_R5G6B5 IF_GUID_EQUAL_RETURN(D3DFMT_R5G6B5); #endif #ifdef D3DFMT_P8 IF_GUID_EQUAL_RETURN(D3DFMT_P8); #endif #ifdef D3DFMT_A2R10G10B10 IF_GUID_EQUAL_RETURN(D3DFMT_A2R10G10B10); #endif #ifdef D3DFMT_A2B10G10R10 IF_GUID_EQUAL_RETURN(D3DFMT_A2B10G10R10); #endif #ifdef D3DFMT_L8 IF_GUID_EQUAL_RETURN(D3DFMT_L8); #endif #ifdef D3DFMT_L16 IF_GUID_EQUAL_RETURN(D3DFMT_L16); #endif #ifdef D3DFMT_D16 IF_GUID_EQUAL_RETURN(D3DFMT_D16); #endif #ifdef MFVideoFormat_A2R10G10B10 IF_GUID_EQUAL_RETURN(MFVideoFormat_A2R10G10B10); #endif #ifdef MFVideoFormat_A16B16G16R16F IF_GUID_EQUAL_RETURN(MFVideoFormat_A16B16G16R16F); #endif IF_GUID_EQUAL_RETURN(MFAudioFormat_PCM); // WAVE_FORMAT_PCM IF_GUID_EQUAL_RETURN(MFAudioFormat_Float); // WAVE_FORMAT_IEEE_FLOAT IF_GUID_EQUAL_RETURN(MFAudioFormat_DTS); // WAVE_FORMAT_DTS IF_GUID_EQUAL_RETURN(MFAudioFormat_Dolby_AC3_SPDIF); // WAVE_FORMAT_DOLBY_AC3_SPDIF IF_GUID_EQUAL_RETURN(MFAudioFormat_DRM); // WAVE_FORMAT_DRM IF_GUID_EQUAL_RETURN(MFAudioFormat_WMAudioV8); // WAVE_FORMAT_WMAUDIO2 IF_GUID_EQUAL_RETURN(MFAudioFormat_WMAudioV9); // WAVE_FORMAT_WMAUDIO3 IF_GUID_EQUAL_RETURN(MFAudioFormat_WMAudio_Lossless); // WAVE_FORMAT_WMAUDIO_LOSSLESS IF_GUID_EQUAL_RETURN(MFAudioFormat_WMASPDIF); // WAVE_FORMAT_WMASPDIF IF_GUID_EQUAL_RETURN(MFAudioFormat_MSP1); // WAVE_FORMAT_WMAVOICE9 IF_GUID_EQUAL_RETURN(MFAudioFormat_MP3); // WAVE_FORMAT_MPEGLAYER3 IF_GUID_EQUAL_RETURN(MFAudioFormat_MPEG); // WAVE_FORMAT_MPEG IF_GUID_EQUAL_RETURN(MFAudioFormat_AAC); // WAVE_FORMAT_MPEG_HEAAC IF_GUID_EQUAL_RETURN(MFAudioFormat_ADTS); // WAVE_FORMAT_MPEG_ADTS_AAC #ifdef MFAudioFormat_ALAC IF_GUID_EQUAL_RETURN(MFAudioFormat_ALAC); #endif #ifdef MFAudioFormat_AMR_NB IF_GUID_EQUAL_RETURN(MFAudioFormat_AMR_NB); #endif #ifdef MFAudioFormat_AMR_WB IF_GUID_EQUAL_RETURN(MFAudioFormat_AMR_WB); #endif #ifdef MFAudioFormat_AMR_WP IF_GUID_EQUAL_RETURN(MFAudioFormat_AMR_WP); #endif #ifdef MFAudioFormat_Dolby_AC3 IF_GUID_EQUAL_RETURN(MFAudioFormat_Dolby_AC3); #endif #ifdef MFAudioFormat_Dolby_DDPlus IF_GUID_EQUAL_RETURN(MFAudioFormat_Dolby_DDPlus); #endif #ifdef MFAudioFormat_FLAC IF_GUID_EQUAL_RETURN(MFAudioFormat_FLAC); #endif #ifdef MFAudioFormat_Opus IF_GUID_EQUAL_RETURN(MFAudioFormat_Opus); #endif #ifdef MEDIASUBTYPE_RAW_AAC1 IF_GUID_EQUAL_RETURN(MEDIASUBTYPE_RAW_AAC1); #endif #ifdef MFAudioFormat_Float_SpatialObjects IF_GUID_EQUAL_RETURN(MFAudioFormat_Float_SpatialObjects); #endif #ifdef MFAudioFormat_QCELP IF_GUID_EQUAL_RETURN(MFAudioFormat_QCELP); #endif return NULL; } bool LogAttributeValueByIndexNew(IMFAttributes *pAttr, DWORD index, MediaType &out) { PROPVARIANT var; PropVariantInit(&var); GUID guid = { 0 }; if (SUCCEEDED(pAttr->GetItemByIndex(index, &guid, &var))) { if (guid == MF_MT_DEFAULT_STRIDE && var.vt == VT_INT) out.MF_MT_DEFAULT_STRIDE = var.intVal; else if (guid == MF_MT_FRAME_RATE && var.vt == VT_UI8) Unpack2UINT32AsUINT64(var.uhVal.QuadPart, &out.MF_MT_FRAME_RATE_NUMERATOR, &out.MF_MT_FRAME_RATE_DENOMINATOR); else if (guid == MF_MT_FRAME_RATE_RANGE_MAX && var.vt == VT_UI8) Unpack2UINT32AsUINT64(var.uhVal.QuadPart, &out.MF_MT_FRAME_RATE_RANGE_MAX_NUMERATOR, &out.MF_MT_FRAME_RATE_RANGE_MAX_DENOMINATOR); else if (guid == MF_MT_FRAME_RATE_RANGE_MIN && var.vt == VT_UI8) Unpack2UINT32AsUINT64(var.uhVal.QuadPart, &out.MF_MT_FRAME_RATE_RANGE_MIN_NUMERATOR, &out.MF_MT_FRAME_RATE_RANGE_MIN_DENOMINATOR); else if (guid == MF_MT_PIXEL_ASPECT_RATIO && var.vt == VT_UI8) Unpack2UINT32AsUINT64(var.uhVal.QuadPart, &out.MF_MT_PIXEL_ASPECT_RATIO_NUMERATOR, &out.MF_MT_PIXEL_ASPECT_RATIO_DENOMINATOR); else if (guid == MF_MT_YUV_MATRIX && var.vt == VT_UI4) out.MF_MT_YUV_MATRIX = var.ulVal; else if (guid == MF_MT_VIDEO_LIGHTING && var.vt == VT_UI4) out.MF_MT_VIDEO_LIGHTING = var.ulVal; else if (guid == MF_MT_DEFAULT_STRIDE && var.vt == VT_UI4) out.MF_MT_DEFAULT_STRIDE = (int)var.ulVal; else if (guid == MF_MT_VIDEO_CHROMA_SITING && var.vt == VT_UI4) out.MF_MT_VIDEO_CHROMA_SITING = var.ulVal; else if (guid == MF_MT_VIDEO_NOMINAL_RANGE && var.vt == VT_UI4) out.MF_MT_VIDEO_NOMINAL_RANGE = var.ulVal; else if (guid == MF_MT_ALL_SAMPLES_INDEPENDENT && var.vt == VT_UI4) out.MF_MT_ALL_SAMPLES_INDEPENDENT = var.ulVal; else if (guid == MF_MT_FIXED_SIZE_SAMPLES && var.vt == VT_UI4) out.MF_MT_FIXED_SIZE_SAMPLES = var.ulVal; else if (guid == MF_MT_SAMPLE_SIZE && var.vt == VT_UI4) out.MF_MT_SAMPLE_SIZE = var.ulVal; else if (guid == MF_MT_VIDEO_PRIMARIES && var.vt == VT_UI4) out.MF_MT_VIDEO_PRIMARIES = var.ulVal; else if (guid == MF_MT_INTERLACE_MODE && var.vt == VT_UI4) out.MF_MT_INTERLACE_MODE = var.ulVal; else if (guid == MF_MT_AM_FORMAT_TYPE && var.vt == VT_CLSID) out.MF_MT_AM_FORMAT_TYPE = *var.puuid; else if (guid == MF_MT_MAJOR_TYPE && var.vt == VT_CLSID) out.pMF_MT_MAJOR_TYPEName = GetGUIDNameConstNew(out.MF_MT_MAJOR_TYPE = *var.puuid); else if (guid == MF_MT_SUBTYPE && var.vt == VT_CLSID) out.pMF_MT_SUBTYPEName = GetGUIDNameConstNew(out.MF_MT_SUBTYPE = *var.puuid); else if (guid == MF_MT_FRAME_SIZE && var.vt == VT_UI8) { Unpack2UINT32AsUINT64(var.uhVal.QuadPart, &out.width, &out.height); out.MF_MT_FRAME_SIZE = out.width * out.height; } PropVariantClear(&var); return true; } return false; } MediaType::MediaType() { pMF_MT_MAJOR_TYPEName = NULL; pMF_MT_SUBTYPEName = NULL; Clear(); } MediaType::MediaType(IMFMediaType *pType) { pMF_MT_MAJOR_TYPEName = NULL; pMF_MT_SUBTYPEName = NULL; Clear(); UINT32 count = 0; if (SUCCEEDED(pType->GetCount(&count)) && SUCCEEDED(pType->LockStore())) { for (UINT32 i = 0; i < count; i++) if (!LogAttributeValueByIndexNew(pType, i, *this)) break; pType->UnlockStore(); } } MediaType::~MediaType() { Clear(); } void MediaType::Clear() { MF_MT_FRAME_SIZE = 0; height = 0; width = 0; MF_MT_YUV_MATRIX = 0; MF_MT_VIDEO_LIGHTING = 0; MF_MT_DEFAULT_STRIDE = 0; MF_MT_VIDEO_CHROMA_SITING = 0; MF_MT_FIXED_SIZE_SAMPLES = 0; MF_MT_VIDEO_NOMINAL_RANGE = 0; MF_MT_FRAME_RATE_NUMERATOR = 0; MF_MT_FRAME_RATE_DENOMINATOR = 0; MF_MT_PIXEL_ASPECT_RATIO_NUMERATOR = 0; MF_MT_PIXEL_ASPECT_RATIO_DENOMINATOR = 0; MF_MT_ALL_SAMPLES_INDEPENDENT = 0; MF_MT_FRAME_RATE_RANGE_MIN_NUMERATOR = 0; MF_MT_FRAME_RATE_RANGE_MIN_DENOMINATOR = 0; MF_MT_SAMPLE_SIZE = 0; MF_MT_VIDEO_PRIMARIES = 0; MF_MT_INTERLACE_MODE = 0; MF_MT_FRAME_RATE_RANGE_MAX_NUMERATOR = 0; MF_MT_FRAME_RATE_RANGE_MAX_DENOMINATOR = 0; memset(&MF_MT_MAJOR_TYPE, 0, sizeof(GUID)); memset(&MF_MT_AM_FORMAT_TYPE, 0, sizeof(GUID)); memset(&MF_MT_SUBTYPE, 0, sizeof(GUID)); } } class SourceReaderCB : public IMFSourceReaderCallback { public: SourceReaderCB() : m_nRefCount(1), m_hEvent(CreateEvent(NULL, FALSE, FALSE, NULL)), m_bEOS(FALSE), m_hrStatus(S_OK), m_dwStreamIndex(0) { } // IUnknown methods STDMETHODIMP QueryInterface(REFIID iid, void** ppv) CV_OVERRIDE { #ifdef _MSC_VER #pragma warning(push) #pragma warning(disable:4838) #endif static const QITAB qit[] = { QITABENT(SourceReaderCB, IMFSourceReaderCallback), { 0 }, }; #ifdef _MSC_VER #pragma warning(pop) #endif return QISearch(this, qit, iid, ppv); } STDMETHODIMP_(ULONG) AddRef() CV_OVERRIDE { return InterlockedIncrement(&m_nRefCount); } STDMETHODIMP_(ULONG) Release() CV_OVERRIDE { ULONG uCount = InterlockedDecrement(&m_nRefCount); if (uCount == 0) { delete this; } return uCount; } STDMETHODIMP OnReadSample(HRESULT hrStatus, DWORD dwStreamIndex, DWORD dwStreamFlags, LONGLONG llTimestamp, IMFSample *pSample) CV_OVERRIDE; STDMETHODIMP OnEvent(DWORD, IMFMediaEvent *) CV_OVERRIDE { return S_OK; } STDMETHODIMP OnFlush(DWORD) CV_OVERRIDE { return S_OK; } HRESULT Wait(DWORD dwMilliseconds, _ComPtr& videoSample, BOOL& pbEOS); private: // Destructor is private. Caller should call Release. virtual ~SourceReaderCB() { CV_LOG_WARNING(NULL, "terminating async callback"); } public: long m_nRefCount; // Reference count. cv::Mutex m_mutex; HANDLE m_hEvent; BOOL m_bEOS; HRESULT m_hrStatus; _ComPtr m_reader; DWORD m_dwStreamIndex; _ComPtr m_lastSample; }; /******* Capturing video from camera or file via Microsoft Media Foundation **********/ class CvCapture_MSMF : public cv::IVideoCapture { public: typedef enum { MODE_SW = 0, MODE_HW = 1 } MSMFCapture_Mode; CvCapture_MSMF(); CvCapture_MSMF(int); CvCapture_MSMF(const cv::String&); virtual ~CvCapture_MSMF(); virtual bool open(int); virtual bool open(const cv::String&); virtual void close(); virtual double getProperty(int) const CV_OVERRIDE; virtual bool setProperty(int, double) CV_OVERRIDE; virtual bool grabFrame() CV_OVERRIDE; virtual bool retrieveFrame(int, cv::OutputArray) CV_OVERRIDE; virtual bool isOpened() const CV_OVERRIDE { return isOpen; } virtual int getCaptureDomain() CV_OVERRIDE { return CV_CAP_MSMF; } // Return the type of the capture object: CV_CAP_VFW, etc... protected: double getFramerate(MediaType MT) const; bool configureOutput(UINT32 width, UINT32 height, double prefFramerate, UINT32 aspectRatioN, UINT32 aspectRatioD, int outFormat, bool convertToFormat); bool setTime(double time, bool rough); bool configureHW(bool enable); Media_Foundation& MF; cv::String filename; int camid; MSMFCapture_Mode captureMode; #ifdef HAVE_DXVA _ComPtr D3DDev; _ComPtr D3DMgr; #endif _ComPtr videoFileSource; DWORD dwStreamIndex; MediaType nativeFormat; MediaType captureFormat; int outputFormat; UINT32 requestedWidth, requestedHeight; bool convertFormat; UINT32 aspectN, aspectD; MFTIME duration; LONGLONG frameStep; _ComPtr videoSample; LONGLONG sampleTime; bool isOpen; _ComPtr readCallback; // non-NULL for "live" streams (camera capture) }; CvCapture_MSMF::CvCapture_MSMF(): MF(Media_Foundation::getInstance()), filename(""), camid(-1), captureMode(MODE_SW), #ifdef HAVE_DXVA D3DDev(NULL), D3DMgr(NULL), #endif videoFileSource(NULL), videoSample(NULL), outputFormat(CV_CAP_MODE_BGR), requestedWidth(0), requestedHeight(0), convertFormat(true), aspectN(1), aspectD(1), sampleTime(0), isOpen(false) { configureHW(true); } CvCapture_MSMF::CvCapture_MSMF(int index) : CvCapture_MSMF() { open(index); } CvCapture_MSMF::CvCapture_MSMF(const cv::String& _filename) : CvCapture_MSMF() { open(_filename); } CvCapture_MSMF::~CvCapture_MSMF() { close(); configureHW(false); } void CvCapture_MSMF::close() { if (isOpen) { isOpen = false; videoSample.Release(); videoFileSource.Release(); camid = -1; filename.clear(); } readCallback.Release(); } bool CvCapture_MSMF::configureHW(bool enable) { #ifdef HAVE_DXVA if ((enable && D3DMgr && D3DDev) || (!enable && !D3DMgr && !D3DDev)) return true; bool reopen = isOpen; int prevcam = camid; cv::String prevfile = filename; close(); if (enable) { D3D_FEATURE_LEVEL levels[] = { D3D_FEATURE_LEVEL_11_1, D3D_FEATURE_LEVEL_11_0, D3D_FEATURE_LEVEL_10_1, D3D_FEATURE_LEVEL_10_0, D3D_FEATURE_LEVEL_9_3, D3D_FEATURE_LEVEL_9_2, D3D_FEATURE_LEVEL_9_1 }; if (SUCCEEDED(D3D11CreateDevice(NULL, D3D_DRIVER_TYPE_HARDWARE, NULL, D3D11_CREATE_DEVICE_BGRA_SUPPORT | D3D11_CREATE_DEVICE_VIDEO_SUPPORT, levels, sizeof(levels) / sizeof(*levels), D3D11_SDK_VERSION, &D3DDev, NULL, NULL))) { // NOTE: Getting ready for multi-threaded operation _ComPtr D3DDevMT; UINT mgrRToken; if (SUCCEEDED(D3DDev->QueryInterface(IID_PPV_ARGS(&D3DDevMT)))) { D3DDevMT->SetMultithreadProtected(TRUE); D3DDevMT.Release(); if (SUCCEEDED(MFCreateDXGIDeviceManager(&mgrRToken, &D3DMgr))) { if (SUCCEEDED(D3DMgr->ResetDevice(D3DDev.Get(), mgrRToken))) { captureMode = MODE_HW; return reopen ? (prevcam >= 0 ? open(prevcam) : open(prevfile.c_str())) : true; } D3DMgr.Release(); } } D3DDev.Release(); } return false; } else { if (D3DMgr) D3DMgr.Release(); if (D3DDev) D3DDev.Release(); captureMode = MODE_SW; return reopen ? (prevcam >= 0 ? open(prevcam) : open(prevfile.c_str())) : true; } #else return !enable; #endif } #define UDIFF(res, ref) (ref == 0 ? 0 : res > ref ? res - ref : ref - res) static UINT32 resolutionDiff(MediaType& mType, UINT32 refWidth, UINT32 refHeight) { return UDIFF(mType.width, refWidth) + UDIFF(mType.height, refHeight); } #undef UDIFF bool CvCapture_MSMF::configureOutput(UINT32 width, UINT32 height, double prefFramerate, UINT32 aspectRatioN, UINT32 aspectRatioD, int outFormat, bool convertToFormat) { if (width != 0 && height != 0 && width == captureFormat.width && height == captureFormat.height && prefFramerate == getFramerate(nativeFormat) && aspectRatioN == aspectN && aspectRatioD == aspectD && outFormat == outputFormat && convertToFormat == convertFormat) return true; requestedWidth = width; requestedHeight = height; HRESULT hr = S_OK; int dwStreamBest = -1; MediaType MTBest; DWORD dwMediaTypeTest = 0; DWORD dwStreamTest = 0; while (SUCCEEDED(hr)) { _ComPtr pType; hr = videoFileSource->GetNativeMediaType(dwStreamTest, dwMediaTypeTest, &pType); if (hr == MF_E_NO_MORE_TYPES) { hr = S_OK; ++dwStreamTest; dwMediaTypeTest = 0; } else if (SUCCEEDED(hr)) { MediaType MT(pType.Get()); if (MT.MF_MT_MAJOR_TYPE == MFMediaType_Video) { if (dwStreamBest < 0 || resolutionDiff(MT, width, height) < resolutionDiff(MTBest, width, height) || (resolutionDiff(MT, width, height) == resolutionDiff(MTBest, width, height) && MT.width > MTBest.width) || (resolutionDiff(MT, width, height) == resolutionDiff(MTBest, width, height) && MT.width == MTBest.width && MT.height > MTBest.height) || (MT.width == MTBest.width && MT.height == MTBest.height && (getFramerate(MT) > getFramerate(MTBest) && (prefFramerate == 0 || getFramerate(MT) <= prefFramerate))) ) { dwStreamBest = (int)dwStreamTest; MTBest = MT; } } ++dwMediaTypeTest; } } if (dwStreamBest >= 0) { GUID outSubtype = GUID_NULL; UINT32 outStride = 0; UINT32 outSize = 0; if(convertToFormat) switch (outFormat) { case CV_CAP_MODE_BGR: case CV_CAP_MODE_RGB: outSubtype = captureMode == MODE_HW ? MFVideoFormat_RGB32 : MFVideoFormat_RGB24; // HW accelerated mode support only RGB32 outStride = (captureMode == MODE_HW ? 4 : 3) * MTBest.width; outSize = outStride * MTBest.height; break; case CV_CAP_MODE_GRAY: outSubtype = MFVideoFormat_NV12; outStride = MTBest.width; outSize = outStride * MTBest.height * 3 / 2; break; case CV_CAP_MODE_YUYV: outSubtype = MFVideoFormat_YUY2; outStride = 2 * MTBest.width; outSize = outStride * MTBest.height; break; default: return false; } _ComPtr mediaTypeOut; if (// Set the output media type. SUCCEEDED(MFCreateMediaType(&mediaTypeOut)) && SUCCEEDED(mediaTypeOut->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video)) && SUCCEEDED(mediaTypeOut->SetGUID(MF_MT_SUBTYPE, convertToFormat ? outSubtype : MTBest.MF_MT_SUBTYPE)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_INTERLACE_MODE, convertToFormat ? MFVideoInterlace_Progressive : MTBest.MF_MT_INTERLACE_MODE)) && SUCCEEDED(MFSetAttributeRatio(mediaTypeOut.Get(), MF_MT_PIXEL_ASPECT_RATIO, aspectRatioN, aspectRatioD)) && SUCCEEDED(MFSetAttributeSize(mediaTypeOut.Get(), MF_MT_FRAME_SIZE, MTBest.width, MTBest.height)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_FIXED_SIZE_SAMPLES, convertToFormat ? 1 : MTBest.MF_MT_FIXED_SIZE_SAMPLES)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_SAMPLE_SIZE, convertToFormat ? outSize : MTBest.MF_MT_SAMPLE_SIZE)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_DEFAULT_STRIDE, convertToFormat ? outStride : MTBest.MF_MT_DEFAULT_STRIDE)))//Assume BGR24 input { if (SUCCEEDED(videoFileSource->SetStreamSelection((DWORD)MF_SOURCE_READER_ALL_STREAMS, false)) && SUCCEEDED(videoFileSource->SetStreamSelection((DWORD)dwStreamBest, true)) && SUCCEEDED(videoFileSource->SetCurrentMediaType((DWORD)dwStreamBest, NULL, mediaTypeOut.Get())) ) { dwStreamIndex = (DWORD)dwStreamBest; nativeFormat = MTBest; aspectN = aspectRatioN; aspectD = aspectRatioD; outputFormat = outFormat; convertFormat = convertToFormat; captureFormat = MediaType(mediaTypeOut.Get()); return true; } close(); } } return false; } // Initialize camera input bool CvCapture_MSMF::open(int _index) { close(); if (_index < 0) return false; _ComPtr msAttr = NULL; if (SUCCEEDED(MFCreateAttributes(&msAttr, 1)) && SUCCEEDED(msAttr->SetGUID( MF_DEVSOURCE_ATTRIBUTE_SOURCE_TYPE, MF_DEVSOURCE_ATTRIBUTE_SOURCE_TYPE_VIDCAP_GUID ))) { IMFActivate **ppDevices = NULL; UINT32 count; if (SUCCEEDED(MFEnumDeviceSources(msAttr.Get(), &ppDevices, &count))) { if (count > 0) { for (int ind = 0; ind < (int)count; ind++) { if (ind == _index && ppDevices[ind]) { // Set source reader parameters _ComPtr mSrc; _ComPtr srAttr; if (SUCCEEDED(ppDevices[ind]->ActivateObject(__uuidof(IMFMediaSource), (void**)&mSrc)) && mSrc && SUCCEEDED(MFCreateAttributes(&srAttr, 10)) && SUCCEEDED(srAttr->SetUINT32(MF_READWRITE_ENABLE_HARDWARE_TRANSFORMS, TRUE)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_DISABLE_DXVA, FALSE)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_ENABLE_VIDEO_PROCESSING, FALSE)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_ENABLE_ADVANCED_VIDEO_PROCESSING, TRUE))) { #ifdef HAVE_DXVA if (D3DMgr) srAttr->SetUnknown(MF_SOURCE_READER_D3D_MANAGER, D3DMgr.Get()); #endif readCallback = ComPtr(new SourceReaderCB()); HRESULT hr = srAttr->SetUnknown(MF_SOURCE_READER_ASYNC_CALLBACK, (IMFSourceReaderCallback*)readCallback.Get()); if (FAILED(hr)) { readCallback.Release(); continue; } if (SUCCEEDED(MFCreateSourceReaderFromMediaSource(mSrc.Get(), srAttr.Get(), &videoFileSource))) { isOpen = true; duration = 0; if (configureOutput(640, 480, 0, aspectN, aspectD, outputFormat, convertFormat)) { double fps = getFramerate(nativeFormat); frameStep = (LONGLONG)(fps > 0 ? 1e7 / fps : 0); camid = _index; } } } } if (ppDevices[ind]) ppDevices[ind]->Release(); } } } CoTaskMemFree(ppDevices); } return isOpen; } bool CvCapture_MSMF::open(const cv::String& _filename) { close(); if (_filename.empty()) return false; // Set source reader parameters _ComPtr srAttr; if (SUCCEEDED(MFCreateAttributes(&srAttr, 10)) && SUCCEEDED(srAttr->SetUINT32(MF_READWRITE_ENABLE_HARDWARE_TRANSFORMS, true)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_DISABLE_DXVA, false)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_ENABLE_VIDEO_PROCESSING, false)) && SUCCEEDED(srAttr->SetUINT32(MF_SOURCE_READER_ENABLE_ADVANCED_VIDEO_PROCESSING, true)) ) { #ifdef HAVE_DXVA if(D3DMgr) srAttr->SetUnknown(MF_SOURCE_READER_D3D_MANAGER, D3DMgr.Get()); #endif cv::AutoBuffer unicodeFileName(_filename.length() + 1); MultiByteToWideChar(CP_ACP, 0, _filename.c_str(), -1, unicodeFileName, (int)_filename.length() + 1); if (SUCCEEDED(MFCreateSourceReaderFromURL(unicodeFileName, srAttr.Get(), &videoFileSource))) { isOpen = true; sampleTime = 0; if (configureOutput(0, 0, 0, aspectN, aspectD, outputFormat, convertFormat)) { double fps = getFramerate(nativeFormat); frameStep = (LONGLONG)(fps > 0 ? 1e7 / fps : 0); filename = _filename; PROPVARIANT var; HRESULT hr; if (SUCCEEDED(hr = videoFileSource->GetPresentationAttribute((DWORD)MF_SOURCE_READER_MEDIASOURCE, MF_PD_DURATION, &var)) && var.vt == VT_UI8) { duration = var.uhVal.QuadPart; PropVariantClear(&var); } else duration = 0; } } } return isOpen; } HRESULT SourceReaderCB::Wait(DWORD dwMilliseconds, _ComPtr& videoSample, BOOL& bEOS) { bEOS = FALSE; DWORD dwResult = WaitForSingleObject(m_hEvent, dwMilliseconds); if (dwResult == WAIT_TIMEOUT) { return E_PENDING; } else if (dwResult != WAIT_OBJECT_0) { return HRESULT_FROM_WIN32(GetLastError()); } if (!bEOS) { cv::AutoLock lock(m_mutex); bEOS = m_bEOS; if (!bEOS) { videoSample = m_lastSample; CV_Assert(videoSample); m_lastSample.Release(); ResetEvent(m_hEvent); // event is auto-reset, but we need this forced reset due time gap between wait() and mutex hold. } } return m_hrStatus; } STDMETHODIMP SourceReaderCB::OnReadSample(HRESULT hrStatus, DWORD dwStreamIndex, DWORD dwStreamFlags, LONGLONG llTimestamp, IMFSample *pSample) { CV_UNUSED(llTimestamp); HRESULT hr = 0; cv::AutoLock lock(m_mutex); if (SUCCEEDED(hrStatus)) { if (pSample) { CV_LOG_DEBUG(NULL, "videoio(MSMF): got frame at " << llTimestamp); IMFSample* prev = m_lastSample.Get(); if (prev) { CV_LOG_DEBUG(NULL, "videoio(MSMF): drop frame (not processed)"); } m_lastSample = pSample; } } else { CV_LOG_WARNING(NULL, "videoio(MSMF): OnReadSample() is called with error status: " << hrStatus); } if (MF_SOURCE_READERF_ENDOFSTREAM & dwStreamFlags) { // Reached the end of the stream. m_bEOS = true; } m_hrStatus = hrStatus; if (FAILED(hr = m_reader->ReadSample(dwStreamIndex, 0, NULL, NULL, NULL, NULL))) { CV_LOG_WARNING(NULL, "videoio(MSMF): async ReadSample() call is failed with error status: " << hr); m_bEOS = true; } if (pSample || m_bEOS) { SetEvent(m_hEvent); } return S_OK; } bool CvCapture_MSMF::grabFrame() { CV_TRACE_FUNCTION(); if (readCallback) // async "live" capture mode { HRESULT hr = 0; SourceReaderCB* reader = ((SourceReaderCB*)readCallback.Get()); if (!reader->m_reader) { // Initiate capturing with async callback reader->m_reader = videoFileSource; reader->m_dwStreamIndex = dwStreamIndex; if (FAILED(hr = videoFileSource->ReadSample(dwStreamIndex, 0, NULL, NULL, NULL, NULL))) { CV_LOG_ERROR(NULL, "videoio(MSMF): can't grab frame - initial async ReadSample() call failed: " << hr); reader->m_reader = NULL; return false; } } BOOL bEOS = false; if (FAILED(hr = reader->Wait(10000, videoSample, bEOS))) // 10 sec { CV_LOG_WARNING(NULL, "videoio(MSMF): can't grab frame. Error: " << hr); return false; } if (bEOS) { CV_LOG_WARNING(NULL, "videoio(MSMF): EOS signal. Capture stream is lost"); return false; } return true; } else if (isOpen) { DWORD streamIndex, flags; videoSample.Release(); HRESULT hr; for(;;) { CV_TRACE_REGION("ReadSample"); if (!SUCCEEDED(hr = videoFileSource->ReadSample( dwStreamIndex, // Stream index. 0, // Flags. &streamIndex, // Receives the actual stream index. &flags, // Receives status flags. &sampleTime, // Receives the time stamp. &videoSample // Receives the sample or NULL. ))) break; if (streamIndex != dwStreamIndex) break; if (flags & (MF_SOURCE_READERF_ERROR | MF_SOURCE_READERF_ALLEFFECTSREMOVED | MF_SOURCE_READERF_ENDOFSTREAM)) break; if (videoSample) break; if (flags & MF_SOURCE_READERF_STREAMTICK) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Stream tick detected. Retrying to grab the frame"); } } if (SUCCEEDED(hr)) { if (streamIndex != dwStreamIndex) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Wrong stream readed. Abort capturing"); close(); } else if (flags & MF_SOURCE_READERF_ERROR) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Stream reading error. Abort capturing"); close(); } else if (flags & MF_SOURCE_READERF_ALLEFFECTSREMOVED) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Stream decoding error. Abort capturing"); close(); } else if (flags & MF_SOURCE_READERF_ENDOFSTREAM) { sampleTime += frameStep; CV_LOG_DEBUG(NULL, "videoio(MSMF): End of stream detected"); } else { sampleTime += frameStep; if (flags & MF_SOURCE_READERF_NEWSTREAM) { CV_LOG_DEBUG(NULL, "videoio(MSMF): New stream detected"); } if (flags & MF_SOURCE_READERF_NATIVEMEDIATYPECHANGED) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Stream native media type changed"); } if (flags & MF_SOURCE_READERF_CURRENTMEDIATYPECHANGED) { CV_LOG_DEBUG(NULL, "videoio(MSMF): Stream current media type changed"); } return true; } } } return false; } bool CvCapture_MSMF::retrieveFrame(int, cv::OutputArray frame) { CV_TRACE_FUNCTION(); do { if (!videoSample) break; _ComPtr buf = NULL; CV_TRACE_REGION("get_contiguous_buffer"); if (!SUCCEEDED(videoSample->ConvertToContiguousBuffer(&buf))) { CV_TRACE_REGION("get_buffer"); DWORD bcnt = 0; if (!SUCCEEDED(videoSample->GetBufferCount(&bcnt))) break; if (bcnt == 0) break; if (!SUCCEEDED(videoSample->GetBufferByIndex(0, &buf))) break; } bool lock2d = false; BYTE* ptr = NULL; LONG pitch = 0; DWORD maxsize = 0, cursize = 0; // "For 2-D buffers, the Lock2D method is more efficient than the Lock method" // see IMFMediaBuffer::Lock method documentation: https://msdn.microsoft.com/en-us/library/windows/desktop/bb970366(v=vs.85).aspx _ComPtr buffer2d; if (convertFormat) { if (SUCCEEDED(buf.As(buffer2d))) { CV_TRACE_REGION_NEXT("lock2d"); if (SUCCEEDED(buffer2d->Lock2D(&ptr, &pitch))) { lock2d = true; } } } if (ptr == NULL) { CV_Assert(lock2d == false); CV_TRACE_REGION_NEXT("lock"); if (!SUCCEEDED(buf->Lock(&ptr, &maxsize, &cursize))) { break; } } if (!ptr) break; if (convertFormat) { if (lock2d || (unsigned int)cursize == captureFormat.MF_MT_SAMPLE_SIZE) { switch (outputFormat) { case CV_CAP_MODE_YUYV: cv::Mat(captureFormat.height, captureFormat.width, CV_8UC2, ptr, pitch).copyTo(frame); break; case CV_CAP_MODE_BGR: if (captureMode == MODE_HW) cv::cvtColor(cv::Mat(captureFormat.height, captureFormat.width, CV_8UC4, ptr, pitch), frame, cv::COLOR_BGRA2BGR); else cv::Mat(captureFormat.height, captureFormat.width, CV_8UC3, ptr, pitch).copyTo(frame); break; case CV_CAP_MODE_RGB: if (captureMode == MODE_HW) cv::cvtColor(cv::Mat(captureFormat.height, captureFormat.width, CV_8UC4, ptr, pitch), frame, cv::COLOR_BGRA2BGR); else cv::cvtColor(cv::Mat(captureFormat.height, captureFormat.width, CV_8UC3, ptr, pitch), frame, cv::COLOR_BGR2RGB); break; case CV_CAP_MODE_GRAY: cv::Mat(captureFormat.height, captureFormat.width, CV_8UC1, ptr, pitch).copyTo(frame); break; default: frame.release(); break; } } else frame.release(); } else { cv::Mat(1, cursize, CV_8UC1, ptr, pitch).copyTo(frame); } CV_TRACE_REGION_NEXT("unlock"); if (lock2d) buffer2d->Unlock2D(); else buf->Unlock(); return !frame.empty(); } while (0); frame.release(); return false; } double CvCapture_MSMF::getFramerate(MediaType MT) const { if (MT.MF_MT_SUBTYPE == MFVideoFormat_MP43) //Unable to estimate FPS for MP43 return 0; return MT.MF_MT_FRAME_RATE_DENOMINATOR != 0 ? ((double)MT.MF_MT_FRAME_RATE_NUMERATOR) / ((double)MT.MF_MT_FRAME_RATE_DENOMINATOR) : 0; } bool CvCapture_MSMF::setTime(double time, bool rough) { PROPVARIANT var; if (SUCCEEDED(videoFileSource->GetPresentationAttribute((DWORD)MF_SOURCE_READER_MEDIASOURCE, MF_SOURCE_READER_MEDIASOURCE_CHARACTERISTICS, &var)) && var.vt == VT_UI4 && var.ulVal & MFMEDIASOURCE_CAN_SEEK) { videoSample.Release(); bool useGrabbing = time > 0 && !rough && !(var.ulVal & MFMEDIASOURCE_HAS_SLOW_SEEK); PropVariantClear(&var); sampleTime = (useGrabbing && time >= frameStep) ? (LONGLONG)floor(time + 0.5) - frameStep : (LONGLONG)floor(time + 0.5); var.vt = VT_I8; var.hVal.QuadPart = sampleTime; bool resOK = SUCCEEDED(videoFileSource->SetCurrentPosition(GUID_NULL, var)); PropVariantClear(&var); if (resOK && useGrabbing) { LONGLONG timeborder = (LONGLONG)floor(time + 0.5) - frameStep / 2; do { resOK = grabFrame(); videoSample.Release(); } while (resOK && sampleTime < timeborder); } return resOK; } return false; } double CvCapture_MSMF::getProperty( int property_id ) const { IAMVideoProcAmp *pProcAmp = NULL; IAMCameraControl *pProcControl = NULL; // image format properties if (isOpen) switch (property_id) { case CV_CAP_PROP_FORMAT: return outputFormat; case CV_CAP_PROP_MODE: return captureMode; case CV_CAP_PROP_CONVERT_RGB: return convertFormat ? 1 : 0; case CV_CAP_PROP_SAR_NUM: return aspectN; case CV_CAP_PROP_SAR_DEN: return aspectD; case CV_CAP_PROP_FRAME_WIDTH: return captureFormat.width; case CV_CAP_PROP_FRAME_HEIGHT: return captureFormat.height; case CV_CAP_PROP_FOURCC: return nativeFormat.MF_MT_SUBTYPE.Data1; case CV_CAP_PROP_FPS: return getFramerate(nativeFormat); case CV_CAP_PROP_FRAME_COUNT: if (duration != 0) return floor(((double)duration / 1e7)*getFramerate(nativeFormat) + 0.5); else break; case CV_CAP_PROP_POS_FRAMES: return floor(((double)sampleTime / 1e7)*getFramerate(nativeFormat) + 0.5); case CV_CAP_PROP_POS_MSEC: return (double)sampleTime / 1e4; case CV_CAP_PROP_POS_AVI_RATIO: if (duration != 0) return (double)sampleTime / duration; else break; case CV_CAP_PROP_BRIGHTNESS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Brightness, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if(FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Brightness, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_CONTRAST: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Contrast, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Contrast, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_SATURATION: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Saturation, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Saturation, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_HUE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Hue, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Hue, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_GAIN: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Gain, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Gain, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_SHARPNESS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Sharpness, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Sharpness, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_GAMMA: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_Gamma, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_Gamma, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_BACKLIGHT: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_BacklightCompensation, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_BacklightCompensation, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_MONOCHROME: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_ColorEnable, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_ColorEnable, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal == 0 ? 1 : 0; } break; case CV_CAP_PROP_TEMPERATURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal, paramFlag; HRESULT hr = pProcAmp->Get(VideoProcAmp_WhiteBalance, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcAmp->GetRange(VideoProcAmp_WhiteBalance, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcAmp->Release(); if (SUCCEEDED(hr)) return paramVal; } case CV_CAP_PROP_WHITE_BALANCE_BLUE_U: case CV_CAP_PROP_WHITE_BALANCE_RED_V: break; case CV_CAP_PROP_PAN: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Pan, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Pan, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_TILT: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Tilt, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Tilt, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_ROLL: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Roll, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Roll, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_IRIS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Iris, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Iris, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_EXPOSURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Exposure, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Exposure, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } case CV_CAP_PROP_AUTO_EXPOSURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Exposure, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Exposure, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramFlag == VideoProcAmp_Flags_Auto; } break; case CV_CAP_PROP_ZOOM: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Zoom, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Zoom, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } break; case CV_CAP_PROP_FOCUS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Focus, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Focus, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramVal; } case CV_CAP_PROP_AUTOFOCUS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal, paramFlag; HRESULT hr = pProcControl->Get(CameraControl_Focus, ¶mVal, ¶mFlag); long minVal, maxVal, stepVal; if (FAILED(hr)) hr = pProcControl->GetRange(CameraControl_Focus, &minVal, &maxVal, &stepVal, ¶mVal, ¶mFlag);//Unable to get the property, trying to return default value pProcControl->Release(); if (SUCCEEDED(hr)) return paramFlag == VideoProcAmp_Flags_Auto; } break; case CV_CAP_PROP_RECTIFICATION: case CV_CAP_PROP_TRIGGER: case CV_CAP_PROP_TRIGGER_DELAY: case CV_CAP_PROP_GUID: case CV_CAP_PROP_ISO_SPEED: case CV_CAP_PROP_SETTINGS: case CV_CAP_PROP_BUFFERSIZE: default: break; } return -1; } bool CvCapture_MSMF::setProperty( int property_id, double value ) { IAMVideoProcAmp *pProcAmp = NULL; IAMCameraControl *pProcControl = NULL; // image capture properties if (isOpen) switch (property_id) { case CV_CAP_PROP_MODE: switch ((MSMFCapture_Mode)((int)value)) { case MODE_SW: return configureHW(false); case MODE_HW: return configureHW(true); default: return false; } case CV_CAP_PROP_FORMAT: return configureOutput(requestedWidth, requestedHeight, getFramerate(nativeFormat), aspectN, aspectD, (int)cvRound(value), convertFormat); case CV_CAP_PROP_CONVERT_RGB: return configureOutput(requestedWidth, requestedHeight, getFramerate(nativeFormat), aspectN, aspectD, outputFormat, value != 0); case CV_CAP_PROP_SAR_NUM: if (value > 0) return configureOutput(requestedWidth, requestedHeight, getFramerate(nativeFormat), (UINT32)cvRound(value), aspectD, outputFormat, convertFormat); break; case CV_CAP_PROP_SAR_DEN: if (value > 0) return configureOutput(requestedWidth, requestedHeight, getFramerate(nativeFormat), aspectN, (UINT32)cvRound(value), outputFormat, convertFormat); break; case CV_CAP_PROP_FRAME_WIDTH: if (value >= 0) return configureOutput((UINT32)cvRound(value), requestedHeight, getFramerate(nativeFormat), aspectN, aspectD, outputFormat, convertFormat); break; case CV_CAP_PROP_FRAME_HEIGHT: if (value >= 0) return configureOutput(requestedWidth, (UINT32)cvRound(value), getFramerate(nativeFormat), aspectN, aspectD, outputFormat, convertFormat); break; case CV_CAP_PROP_FPS: if (value >= 0) return configureOutput(requestedWidth, requestedHeight, value, aspectN, aspectD, outputFormat, convertFormat); break; case CV_CAP_PROP_FOURCC: break; case CV_CAP_PROP_FRAME_COUNT: break; case CV_CAP_PROP_POS_AVI_RATIO: if (duration != 0) return setTime(duration * value, true); break; case CV_CAP_PROP_POS_FRAMES: if (getFramerate(nativeFormat) != 0) return setTime(value * 1e7 / getFramerate(nativeFormat), false); break; case CV_CAP_PROP_POS_MSEC: return setTime(value * 1e4, false); case CV_CAP_PROP_BRIGHTNESS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Brightness, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_CONTRAST: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Contrast, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_SATURATION: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Saturation, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_HUE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Hue, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_GAIN: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Gain, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_SHARPNESS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Sharpness, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_GAMMA: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_Gamma, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_BACKLIGHT: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_BacklightCompensation, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_MONOCHROME: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = value != 0 ? 0 : 1; HRESULT hr = pProcAmp->Set(VideoProcAmp_ColorEnable, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_TEMPERATURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcAmp)))) { long paramVal = (long)value; HRESULT hr = pProcAmp->Set(VideoProcAmp_WhiteBalance, paramVal, VideoProcAmp_Flags_Manual); pProcAmp->Release(); return SUCCEEDED(hr); } case CV_CAP_PROP_WHITE_BALANCE_BLUE_U: case CV_CAP_PROP_WHITE_BALANCE_RED_V: break; case CV_CAP_PROP_PAN: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Pan, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_TILT: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Tilt, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_ROLL: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Roll, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_IRIS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Iris, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_EXPOSURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Exposure, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } case CV_CAP_PROP_AUTO_EXPOSURE: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = 0; HRESULT hr = pProcControl->Set(CameraControl_Exposure, paramVal, value != 0 ? VideoProcAmp_Flags_Auto : VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_ZOOM: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Zoom, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_FOCUS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = (long)value; HRESULT hr = pProcControl->Set(CameraControl_Focus, paramVal, VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } case CV_CAP_PROP_AUTOFOCUS: if (SUCCEEDED(videoFileSource->GetServiceForStream((DWORD)MF_SOURCE_READER_MEDIASOURCE, GUID_NULL, IID_PPV_ARGS(&pProcControl)))) { long paramVal = 0; HRESULT hr = pProcControl->Set(CameraControl_Focus, paramVal, value != 0 ? VideoProcAmp_Flags_Auto : VideoProcAmp_Flags_Manual); pProcControl->Release(); return SUCCEEDED(hr); } break; case CV_CAP_PROP_RECTIFICATION: case CV_CAP_PROP_TRIGGER: case CV_CAP_PROP_TRIGGER_DELAY: case CV_CAP_PROP_GUID: case CV_CAP_PROP_ISO_SPEED: case CV_CAP_PROP_SETTINGS: case CV_CAP_PROP_BUFFERSIZE: default: break; } return false; } cv::Ptr cv::cvCreateCapture_MSMF( int index ) { cv::Ptr capture = cv::makePtr(index); if (capture && capture->isOpened()) return capture; return cv::Ptr(); } cv::Ptr cv::cvCreateCapture_MSMF (const cv::String& filename) { cv::Ptr capture = cv::makePtr(filename); if (capture && capture->isOpened()) return capture; return cv::Ptr(); } // // // Media Foundation-based Video Writer // // class CvVideoWriter_MSMF : public cv::IVideoWriter { public: CvVideoWriter_MSMF(); CvVideoWriter_MSMF(const cv::String& filename, int fourcc, double fps, cv::Size frameSize, bool isColor); virtual ~CvVideoWriter_MSMF(); virtual bool open(const cv::String& filename, int fourcc, double fps, cv::Size frameSize, bool isColor); virtual void close(); virtual void write(cv::InputArray); virtual double getProperty(int) const { return 0; } virtual bool setProperty(int, double) { return false; } virtual bool isOpened() const { return initiated; } private: Media_Foundation& MF; UINT32 videoWidth; UINT32 videoHeight; double fps; UINT32 bitRate; UINT32 frameSize; GUID encodingFormat; GUID inputFormat; DWORD streamIndex; _ComPtr sinkWriter; bool initiated; LONGLONG rtStart; UINT64 rtDuration; static const GUID FourCC2GUID(int fourcc); }; CvVideoWriter_MSMF::CvVideoWriter_MSMF(): MF(Media_Foundation::getInstance()), initiated(false) { } CvVideoWriter_MSMF::CvVideoWriter_MSMF(const cv::String& filename, int fourcc, double fps, cv::Size frameSize, bool isColor) : CvVideoWriter_MSMF() { open(filename, fourcc, fps, frameSize, isColor); } CvVideoWriter_MSMF::~CvVideoWriter_MSMF() { close(); } const GUID CvVideoWriter_MSMF::FourCC2GUID(int fourcc) { switch(fourcc) { case CV_FOURCC_MACRO('d', 'v', '2', '5'): return MFVideoFormat_DV25; break; case CV_FOURCC_MACRO('d', 'v', '5', '0'): return MFVideoFormat_DV50; break; case CV_FOURCC_MACRO('d', 'v', 'c', ' '): return MFVideoFormat_DVC; break; case CV_FOURCC_MACRO('d', 'v', 'h', '1'): return MFVideoFormat_DVH1; break; case CV_FOURCC_MACRO('d', 'v', 'h', 'd'): return MFVideoFormat_DVHD; break; case CV_FOURCC_MACRO('d', 'v', 's', 'd'): return MFVideoFormat_DVSD; break; case CV_FOURCC_MACRO('d', 'v', 's', 'l'): return MFVideoFormat_DVSL; break; #if (WINVER >= 0x0602) case CV_FOURCC_MACRO('H', '2', '6', '3'): // Available only for Win 8 target. return MFVideoFormat_H263; break; #endif case CV_FOURCC_MACRO('H', '2', '6', '4'): return MFVideoFormat_H264; break; case CV_FOURCC_MACRO('M', '4', 'S', '2'): return MFVideoFormat_M4S2; break; case CV_FOURCC_MACRO('M', 'J', 'P', 'G'): return MFVideoFormat_MJPG; break; case CV_FOURCC_MACRO('M', 'P', '4', '3'): return MFVideoFormat_MP43; break; case CV_FOURCC_MACRO('M', 'P', '4', 'S'): return MFVideoFormat_MP4S; break; case CV_FOURCC_MACRO('M', 'P', '4', 'V'): return MFVideoFormat_MP4V; break; case CV_FOURCC_MACRO('M', 'P', 'G', '1'): return MFVideoFormat_MPG1; break; case CV_FOURCC_MACRO('M', 'S', 'S', '1'): return MFVideoFormat_MSS1; break; case CV_FOURCC_MACRO('M', 'S', 'S', '2'): return MFVideoFormat_MSS2; break; case CV_FOURCC_MACRO('W', 'M', 'V', '1'): return MFVideoFormat_WMV1; break; case CV_FOURCC_MACRO('W', 'M', 'V', '2'): return MFVideoFormat_WMV2; break; case CV_FOURCC_MACRO('W', 'M', 'V', '3'): return MFVideoFormat_WMV3; break; case CV_FOURCC_MACRO('W', 'V', 'C', '1'): return MFVideoFormat_WVC1; break; default: return MFVideoFormat_H264; } } bool CvVideoWriter_MSMF::open( const cv::String& filename, int fourcc, double _fps, cv::Size _frameSize, bool /*isColor*/ ) { if (initiated) close(); videoWidth = _frameSize.width; videoHeight = _frameSize.height; fps = _fps; bitRate = (UINT32)fps*videoWidth*videoHeight; // 1-bit per pixel encodingFormat = FourCC2GUID(fourcc); inputFormat = MFVideoFormat_RGB32; _ComPtr mediaTypeOut; _ComPtr mediaTypeIn; _ComPtr spAttr; if (// Set the output media type. SUCCEEDED(MFCreateMediaType(&mediaTypeOut)) && SUCCEEDED(mediaTypeOut->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video)) && SUCCEEDED(mediaTypeOut->SetGUID(MF_MT_SUBTYPE, encodingFormat)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_AVG_BITRATE, bitRate)) && SUCCEEDED(mediaTypeOut->SetUINT32(MF_MT_INTERLACE_MODE, MFVideoInterlace_Progressive)) && SUCCEEDED(MFSetAttributeSize(mediaTypeOut.Get(), MF_MT_FRAME_SIZE, videoWidth, videoHeight)) && SUCCEEDED(MFSetAttributeRatio(mediaTypeOut.Get(), MF_MT_FRAME_RATE, (UINT32)fps, 1)) && SUCCEEDED(MFSetAttributeRatio(mediaTypeOut.Get(), MF_MT_PIXEL_ASPECT_RATIO, 1, 1)) && // Set the input media type. SUCCEEDED(MFCreateMediaType(&mediaTypeIn)) && SUCCEEDED(mediaTypeIn->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video)) && SUCCEEDED(mediaTypeIn->SetGUID(MF_MT_SUBTYPE, inputFormat)) && SUCCEEDED(mediaTypeIn->SetUINT32(MF_MT_INTERLACE_MODE, MFVideoInterlace_Progressive)) && SUCCEEDED(mediaTypeIn->SetUINT32(MF_MT_DEFAULT_STRIDE, 4 * videoWidth)) && //Assume BGR32 input SUCCEEDED(MFSetAttributeSize(mediaTypeIn.Get(), MF_MT_FRAME_SIZE, videoWidth, videoHeight)) && SUCCEEDED(MFSetAttributeRatio(mediaTypeIn.Get(), MF_MT_FRAME_RATE, (UINT32)fps, 1)) && SUCCEEDED(MFSetAttributeRatio(mediaTypeIn.Get(), MF_MT_PIXEL_ASPECT_RATIO, 1, 1)) && // Set sink writer parameters SUCCEEDED(MFCreateAttributes(&spAttr, 10)) && SUCCEEDED(spAttr->SetUINT32(MF_READWRITE_ENABLE_HARDWARE_TRANSFORMS, true)) && SUCCEEDED(spAttr->SetUINT32(MF_SINK_WRITER_DISABLE_THROTTLING, true)) ) { // Create the sink writer cv::AutoBuffer unicodeFileName(filename.length() + 1); MultiByteToWideChar(CP_ACP, 0, filename.c_str(), -1, unicodeFileName, (int)filename.length() + 1); HRESULT hr = MFCreateSinkWriterFromURL(unicodeFileName, NULL, spAttr.Get(), &sinkWriter); if (SUCCEEDED(hr)) { // Configure the sink writer and tell it start to start accepting data if (SUCCEEDED(sinkWriter->AddStream(mediaTypeOut.Get(), &streamIndex)) && SUCCEEDED(sinkWriter->SetInputMediaType(streamIndex, mediaTypeIn.Get(), NULL)) && SUCCEEDED(sinkWriter->BeginWriting())) { initiated = true; rtStart = 0; MFFrameRateToAverageTimePerFrame((UINT32)fps, 1, &rtDuration); return true; } } } return false; } void CvVideoWriter_MSMF::close() { if (initiated) { initiated = false; sinkWriter->Finalize(); sinkWriter.Release(); } } void CvVideoWriter_MSMF::write(cv::InputArray img) { if (img.empty() || (img.channels() != 1 && img.channels() != 3 && img.channels() != 4) || (UINT32)img.cols() != videoWidth || (UINT32)img.rows() != videoHeight) return; const LONG cbWidth = 4 * videoWidth; const DWORD cbBuffer = cbWidth * videoHeight; _ComPtr sample; _ComPtr buffer; BYTE *pData = NULL; // Prepare a media sample. if (SUCCEEDED(MFCreateSample(&sample)) && // Set sample time stamp and duration. SUCCEEDED(sample->SetSampleTime(rtStart)) && SUCCEEDED(sample->SetSampleDuration(rtDuration)) && // Create a memory buffer. SUCCEEDED(MFCreateMemoryBuffer(cbBuffer, &buffer)) && // Set the data length of the buffer. SUCCEEDED(buffer->SetCurrentLength(cbBuffer)) && // Add the buffer to the sample. SUCCEEDED(sample->AddBuffer(buffer.Get())) && // Lock the buffer. SUCCEEDED(buffer->Lock(&pData, NULL, NULL))) { // Copy the video frame to the buffer. cv::cvtColor(img.getMat(), cv::Mat(videoHeight, videoWidth, CV_8UC4, pData, cbWidth), img.channels() > 1 ? cv::COLOR_BGR2BGRA : cv::COLOR_GRAY2BGRA); buffer->Unlock(); // Send media sample to the Sink Writer. if (SUCCEEDED(sinkWriter->WriteSample(streamIndex, sample.Get()))) { rtStart += rtDuration; } } } cv::Ptr cv::cvCreateVideoWriter_MSMF( const cv::String& filename, int fourcc, double fps, cv::Size frameSize, int isColor ) { cv::Ptr writer = cv::makePtr(filename, fourcc, fps, frameSize, isColor != 0); if (writer && writer->isOpened()) return writer; return cv::Ptr(); } #endif