Merge pull request #19609 from alalek:videoio_fix_gstreamer_alignment_bug

pull/19627/head
Alexander Alekhin 4 years ago
commit 7328600a7a
  1. 200
      modules/videoio/src/cap_gstreamer.cpp
  2. 19
      modules/videoio/test/test_gstreamer.cpp

@ -177,6 +177,20 @@ private:
GSafePtr& operator=(const T*); // = disabled GSafePtr& operator=(const T*); // = disabled
}; };
class ScopeGuardGstMapInfo
{
GstBuffer* buf_;
GstMapInfo* info_;
public:
ScopeGuardGstMapInfo(GstBuffer* buf, GstMapInfo* info)
: buf_(buf), info_(info)
{}
~ScopeGuardGstMapInfo()
{
gst_buffer_unmap(buf_, info_);
}
};
} // namespace } // namespace
/*! /*!
@ -300,7 +314,6 @@ public:
static void newPad(GstElement * /*elem*/, GstPad *pad, gpointer data); static void newPad(GstElement * /*elem*/, GstPad *pad, gpointer data);
protected: protected:
bool determineFrameDims(CV_OUT Size& sz, CV_OUT gint& channels, CV_OUT bool& isOutputByteBuffer);
bool isPipelinePlaying(); bool isPipelinePlaying();
void startPipeline(); void startPipeline();
void stopPipeline(); void stopPipeline();
@ -369,72 +382,68 @@ bool GStreamerCapture::grabFrame()
bool GStreamerCapture::retrieveFrame(int, OutputArray dst) bool GStreamerCapture::retrieveFrame(int, OutputArray dst)
{ {
if (!sample) if (!sample)
{
return false; return false;
Size sz; }
gint channels = 0;
bool isOutputByteBuffer = false;
if (!determineFrameDims(sz, channels, isOutputByteBuffer))
return false;
// gstreamer expects us to handle the memory at this point GstCaps* frame_caps = gst_sample_get_caps(sample); // no lifetime transfer
// so we can just wrap the raw buffer and be done with it if (!frame_caps)
GstBuffer* buf = gst_sample_get_buffer(sample); // no lifetime transfer
if (!buf)
return false;
GstMapInfo info = {};
if (!gst_buffer_map(buf, &info, GST_MAP_READ))
{ {
//something weird went wrong here. abort. abort. CV_LOG_ERROR(NULL, "GStreamer: gst_sample_get_caps() returns NULL");
CV_WARN("Failed to map GStreamer buffer to system memory");
return false; return false;
} }
try if (!GST_CAPS_IS_SIMPLE(frame_caps))
{ {
Mat src; // bail out in no caps
if (isOutputByteBuffer) CV_LOG_ERROR(NULL, "GStreamer: GST_CAPS_IS_SIMPLE(frame_caps) check is failed");
src = Mat(Size(info.size, 1), CV_8UC1, info.data); return false;
else
src = Mat(sz, CV_MAKETYPE(CV_8U, channels), info.data);
CV_Assert(src.isContinuous());
src.copyTo(dst);
} }
catch (...)
GstVideoInfo info = {};
gboolean video_info_res = gst_video_info_from_caps(&info, frame_caps);
if (!video_info_res)
{ {
gst_buffer_unmap(buf, &info); CV_Error(Error::StsError, "GStreamer: gst_video_info_from_caps() is failed. Can't handle unknown layout");
throw;
} }
gst_buffer_unmap(buf, &info);
return true;
}
bool GStreamerCapture::determineFrameDims(Size &sz, gint& channels, bool& isOutputByteBuffer) int frame_width = GST_VIDEO_INFO_WIDTH(&info);
{ int frame_height = GST_VIDEO_INFO_HEIGHT(&info);
GstCaps * frame_caps = gst_sample_get_caps(sample); // no lifetime transfer if (frame_width <= 0 || frame_height <= 0)
{
// bail out in no caps CV_LOG_ERROR(NULL, "GStreamer: Can't query frame size from GStreamer sample");
if (!GST_CAPS_IS_SIMPLE(frame_caps))
return false; return false;
}
GstStructure* structure = gst_caps_get_structure(frame_caps, 0); // no lifetime transfer GstStructure* structure = gst_caps_get_structure(frame_caps, 0); // no lifetime transfer
if (!structure)
// bail out if width or height are 0
if (!gst_structure_get_int(structure, "width", &width)
|| !gst_structure_get_int(structure, "height", &height))
{ {
CV_WARN("Can't query frame size from GStreeamer buffer"); CV_LOG_ERROR(NULL, "GStreamer: Can't query 'structure'-0 from GStreamer sample");
return false; return false;
} }
sz = Size(width, height);
const gchar* name_ = gst_structure_get_name(structure); const gchar* name_ = gst_structure_get_name(structure);
if (!name_) if (!name_)
{
CV_LOG_ERROR(NULL, "GStreamer: Can't query 'name' from GStreamer sample");
return false; return false;
}
std::string name = toLowerCase(std::string(name_)); std::string name = toLowerCase(std::string(name_));
// we support 11 types of data: // gstreamer expects us to handle the memory at this point
// so we can just wrap the raw buffer and be done with it
GstBuffer* buf = gst_sample_get_buffer(sample); // no lifetime transfer
if (!buf)
return false;
GstMapInfo map_info = {};
if (!gst_buffer_map(buf, &map_info, GST_MAP_READ))
{
CV_LOG_ERROR(NULL, "GStreamer: Failed to map GStreamer buffer to system memory");
return false;
}
ScopeGuardGstMapInfo map_guard(buf, &map_info); // call gst_buffer_unmap(buf, &map_info) on scope leave
// we support these types of data:
// video/x-raw, format=BGR -> 8bit, 3 channels // video/x-raw, format=BGR -> 8bit, 3 channels
// video/x-raw, format=GRAY8 -> 8bit, 1 channel // video/x-raw, format=GRAY8 -> 8bit, 1 channel
// video/x-raw, format=UYVY -> 8bit, 2 channel // video/x-raw, format=UYVY -> 8bit, 2 channel
@ -448,50 +457,117 @@ bool GStreamerCapture::determineFrameDims(Size &sz, gint& channels, bool& isOutp
// image/jpeg -> 8bit, mjpeg: buffer_size x 1 x 1 // image/jpeg -> 8bit, mjpeg: buffer_size x 1 x 1
// bayer data is never decoded, the user is responsible for that // bayer data is never decoded, the user is responsible for that
// everything is 8 bit, so we just test the caps for bit depth // everything is 8 bit, so we just test the caps for bit depth
Size sz = Size(frame_width, frame_height);
guint n_planes = GST_VIDEO_INFO_N_PLANES(&info);
if (name == "video/x-raw") if (name == "video/x-raw")
{ {
const gchar* format_ = gst_structure_get_string(structure, "format"); const gchar* format_ = gst_structure_get_string(structure, "format");
if (!format_) if (!format_)
{
CV_LOG_ERROR(NULL, "GStreamer: Can't query 'format' of 'video/x-raw'");
return false; return false;
}
std::string format = toUpperCase(std::string(format_)); std::string format = toUpperCase(std::string(format_));
if (format == "BGR") if (format == "BGR")
{ {
channels = 3; CV_CheckEQ((int)n_planes, 1, "");
size_t step = GST_VIDEO_INFO_PLANE_STRIDE(&info, 0);
CV_CheckGE(step, (size_t)frame_width * 3, "");
Mat src(sz, CV_8UC3, map_info.data + GST_VIDEO_INFO_PLANE_OFFSET(&info, 0), step);
src.copyTo(dst);
return true;
} }
else if (format == "UYVY" || format == "YUY2" || format == "YVYU") else if (format == "GRAY8")
{ {
channels = 2; CV_CheckEQ((int)n_planes, 1, "");
size_t step = GST_VIDEO_INFO_PLANE_STRIDE(&info, 0);
CV_CheckGE(step, (size_t)frame_width, "");
Mat src(sz, CV_8UC1, map_info.data + GST_VIDEO_INFO_PLANE_OFFSET(&info, 0), step);
src.copyTo(dst);
return true;
} }
else if (format == "NV12" || format == "NV21" || format == "YV12" || format == "I420") else if (format == "UYVY" || format == "YUY2" || format == "YVYU")
{ {
channels = 1; CV_CheckEQ((int)n_planes, 1, "");
sz.height = sz.height * 3 / 2; size_t step = GST_VIDEO_INFO_PLANE_STRIDE(&info, 0);
CV_CheckGE(step, (size_t)frame_width * 2, "");
Mat src(sz, CV_8UC2, map_info.data + GST_VIDEO_INFO_PLANE_OFFSET(&info, 0), step);
src.copyTo(dst);
return true;
}
else if (format == "NV12" || format == "NV21")
{
CV_CheckEQ((int)n_planes, 2, "");
size_t stepY = GST_VIDEO_INFO_PLANE_STRIDE(&info, 0);
CV_CheckGE(stepY, (size_t)frame_width, "");
size_t stepUV = GST_VIDEO_INFO_PLANE_STRIDE(&info, 1);
CV_CheckGE(stepUV, (size_t)frame_width, "");
size_t offsetY = GST_VIDEO_INFO_PLANE_OFFSET(&info, 0);
size_t offsetUV = GST_VIDEO_INFO_PLANE_OFFSET(&info, 1);
if (stepY != stepUV || (offsetUV - offsetY) != (stepY * frame_height))
{
dst.create(Size(frame_width, frame_height * 3 / 2), CV_8UC1);
Mat dst_ = dst.getMat();
Mat srcY(sz, CV_8UC1, map_info.data + offsetY, stepY);
Mat srcUV(Size(frame_width, frame_height / 2), CV_8UC1, map_info.data + offsetUV, stepUV);
srcY.copyTo(dst_(Rect(0, 0, frame_width, frame_height)));
srcUV.copyTo(dst_(Rect(0, frame_height, frame_width, frame_height / 2)));
}
else
{
Mat src(Size(frame_width, frame_height * 3 / 2), CV_8UC1, map_info.data + offsetY, stepY);
src.copyTo(dst);
}
return true;
} }
else if (format == "GRAY8") else if (format == "YV12" || format == "I420")
{ {
channels = 1; CV_CheckEQ((int)n_planes, 3, "");
size_t step0 = GST_VIDEO_INFO_PLANE_STRIDE(&info, 0);
CV_CheckGE(step0, (size_t)frame_width, "");
size_t step1 = GST_VIDEO_INFO_PLANE_STRIDE(&info, 1);
CV_CheckGE(step1, (size_t)frame_width / 2, "");
size_t step2 = GST_VIDEO_INFO_PLANE_STRIDE(&info, 2);
CV_CheckGE(step2, (size_t)frame_width / 2, "");
size_t offset0 = GST_VIDEO_INFO_PLANE_OFFSET(&info, 0);
size_t offset1 = GST_VIDEO_INFO_PLANE_OFFSET(&info, 1);
size_t offset2 = GST_VIDEO_INFO_PLANE_OFFSET(&info, 2);
{
dst.create(Size(frame_width, frame_height * 3 / 2), CV_8UC1);
Mat dst_ = dst.getMat();
Mat srcY(sz, CV_8UC1, map_info.data + offset0, step0);
Size sz2(frame_width / 2, frame_height / 2);
Mat src1(sz2, CV_8UC1, map_info.data + offset1, step1);
Mat src2(sz2, CV_8UC1, map_info.data + offset2, step2);
srcY.copyTo(dst_(Rect(0, 0, frame_width, frame_height)));
src1.copyTo(Mat(sz2, CV_8UC1, dst_.ptr<uchar>(frame_height)));
src2.copyTo(Mat(sz2, CV_8UC1, dst_.ptr<uchar>(frame_height) + src1.total()));
}
return true;
} }
else else
{ {
CV_Error_(Error::StsNotImplemented, ("Unsupported GStreamer format: %s", format.c_str())); CV_Error_(Error::StsNotImplemented, ("Unsupported GStreamer 'video/x-raw' format: %s", format.c_str()));
} }
} }
else if (name == "video/x-bayer") else if (name == "video/x-bayer")
{ {
channels = 1; CV_CheckEQ((int)n_planes, 0, "");
Mat src = Mat(sz, CV_8UC1, map_info.data);
src.copyTo(dst);
return true;
} }
else if (name == "image/jpeg") else if (name == "image/jpeg")
{ {
// the correct size will be set once the first frame arrives CV_CheckEQ((int)n_planes, 0, "");
channels = 1; Mat src = Mat(Size(map_info.size, 1), CV_8UC1, map_info.data);
isOutputByteBuffer = true; src.copyTo(dst);
} return true;
else
{
CV_Error_(Error::StsNotImplemented, ("Unsupported GStreamer layer type: %s", name.c_str()));
} }
return true;
CV_Error_(Error::StsNotImplemented, ("Unsupported GStreamer layer type: %s", name.c_str()));
} }
bool GStreamerCapture::isPipelinePlaying() bool GStreamerCapture::isPipelinePlaying()

@ -4,13 +4,12 @@
#include "test_precomp.hpp" #include "test_precomp.hpp"
namespace opencv_test namespace opencv_test { namespace {
{
typedef tuple< string, Size, Size, int > Param; typedef tuple< string, Size, Size, int > Param;
typedef testing::TestWithParam< Param > videoio_gstreamer; typedef testing::TestWithParam< Param > videoio_gstreamer;
TEST_P(videoio_gstreamer, read_write) TEST_P(videoio_gstreamer, read_check)
{ {
if (!videoio_registry::hasBackend(CAP_GSTREAMER)) if (!videoio_registry::hasBackend(CAP_GSTREAMER))
throw SkipTestException("GStreamer backend was not found"); throw SkipTestException("GStreamer backend was not found");
@ -57,7 +56,7 @@ TEST_P(videoio_gstreamer, read_write)
ASSERT_FALSE(cap.isOpened()); ASSERT_FALSE(cap.isOpened());
} }
Param test_data[] = { static const Param test_data[] = {
make_tuple("video/x-raw, format=BGR" , Size(640, 480), Size(640, 480), COLOR_BGR2RGB), make_tuple("video/x-raw, format=BGR" , Size(640, 480), Size(640, 480), COLOR_BGR2RGB),
make_tuple("video/x-raw, format=GRAY8", Size(640, 480), Size(640, 480), COLOR_GRAY2RGB), make_tuple("video/x-raw, format=GRAY8", Size(640, 480), Size(640, 480), COLOR_GRAY2RGB),
make_tuple("video/x-raw, format=UYVY" , Size(640, 480), Size(640, 480), COLOR_YUV2RGB_UYVY), make_tuple("video/x-raw, format=UYVY" , Size(640, 480), Size(640, 480), COLOR_YUV2RGB_UYVY),
@ -68,7 +67,15 @@ Param test_data[] = {
make_tuple("video/x-raw, format=YV12" , Size(640, 480), Size(640, 720), COLOR_YUV2RGB_YV12), make_tuple("video/x-raw, format=YV12" , Size(640, 480), Size(640, 720), COLOR_YUV2RGB_YV12),
make_tuple("video/x-raw, format=I420" , Size(640, 480), Size(640, 720), COLOR_YUV2RGB_I420), make_tuple("video/x-raw, format=I420" , Size(640, 480), Size(640, 720), COLOR_YUV2RGB_I420),
make_tuple("video/x-bayer" , Size(640, 480), Size(640, 480), COLOR_BayerBG2RGB), make_tuple("video/x-bayer" , Size(640, 480), Size(640, 480), COLOR_BayerBG2RGB),
make_tuple("jpegenc ! image/jpeg" , Size(640, 480), Size(640, 480), COLOR_BGR2RGB) make_tuple("jpegenc ! image/jpeg" , Size(640, 480), Size(640, 480), COLOR_BGR2RGB),
// unaligned cases, strides information must be used
make_tuple("video/x-raw, format=BGR" , Size(322, 242), Size(322, 242), COLOR_BGR2RGB),
make_tuple("video/x-raw, format=GRAY8", Size(322, 242), Size(322, 242), COLOR_GRAY2RGB),
make_tuple("video/x-raw, format=NV12" , Size(322, 242), Size(322, 363), COLOR_YUV2RGB_NV12),
make_tuple("video/x-raw, format=NV21" , Size(322, 242), Size(322, 363), COLOR_YUV2RGB_NV21),
make_tuple("video/x-raw, format=YV12" , Size(322, 242), Size(322, 363), COLOR_YUV2RGB_YV12),
make_tuple("video/x-raw, format=I420" , Size(322, 242), Size(322, 363), COLOR_YUV2RGB_I420),
}; };
INSTANTIATE_TEST_CASE_P(videoio, videoio_gstreamer, testing::ValuesIn(test_data)); INSTANTIATE_TEST_CASE_P(videoio, videoio_gstreamer, testing::ValuesIn(test_data));
@ -132,4 +139,4 @@ TEST(videoio_gstreamer, gray16_writing)
EXPECT_EQ(0, remove(temp_file.c_str())); EXPECT_EQ(0, remove(temp_file.c_str()));
} }
} // namespace }} // namespace

Loading…
Cancel
Save