|
|
|
@ -11,8 +11,6 @@ Test for Tensorflow models loading |
|
|
|
|
|
|
|
|
|
#include "test_precomp.hpp" |
|
|
|
|
#include "npy_blob.hpp" |
|
|
|
|
#include <opencv2/core/ocl.hpp> |
|
|
|
|
#include <opencv2/ts/ocl_test.hpp> |
|
|
|
|
|
|
|
|
|
namespace opencv_test |
|
|
|
|
{ |
|
|
|
@ -115,118 +113,85 @@ static void runTensorFlowNet(const std::string& prefix, int targetId = DNN_TARGE |
|
|
|
|
normAssert(target, output, "", l1, lInf); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, conv) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("single_conv"); |
|
|
|
|
runTensorFlowNet("atrous_conv2d_valid"); |
|
|
|
|
runTensorFlowNet("atrous_conv2d_same"); |
|
|
|
|
runTensorFlowNet("depthwise_conv2d"); |
|
|
|
|
} |
|
|
|
|
typedef testing::TestWithParam<DNNTarget> Test_TensorFlow_layers; |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, padding) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, conv) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("padding_same"); |
|
|
|
|
runTensorFlowNet("padding_valid"); |
|
|
|
|
runTensorFlowNet("spatial_padding"); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("single_conv", targetId); |
|
|
|
|
runTensorFlowNet("atrous_conv2d_valid", targetId); |
|
|
|
|
runTensorFlowNet("atrous_conv2d_same", targetId); |
|
|
|
|
runTensorFlowNet("depthwise_conv2d", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, eltwise_add_mul) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, padding) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("eltwise_add_mul"); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("padding_same", targetId); |
|
|
|
|
runTensorFlowNet("padding_valid", targetId); |
|
|
|
|
runTensorFlowNet("spatial_padding", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
OCL_TEST(Test_TensorFlow, eltwise_add_mul) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, eltwise_add_mul) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("eltwise_add_mul", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("eltwise_add_mul", GetParam()); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, pad_and_concat) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, pad_and_concat) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("pad_and_concat"); |
|
|
|
|
runTensorFlowNet("pad_and_concat", GetParam()); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, batch_norm) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, batch_norm) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("batch_norm"); |
|
|
|
|
runTensorFlowNet("fused_batch_norm"); |
|
|
|
|
runTensorFlowNet("batch_norm_text", DNN_TARGET_CPU, true); |
|
|
|
|
runTensorFlowNet("mvn_batch_norm"); |
|
|
|
|
runTensorFlowNet("mvn_batch_norm_1x1"); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("batch_norm", targetId); |
|
|
|
|
runTensorFlowNet("fused_batch_norm", targetId); |
|
|
|
|
runTensorFlowNet("batch_norm_text", targetId, true); |
|
|
|
|
runTensorFlowNet("mvn_batch_norm", targetId); |
|
|
|
|
runTensorFlowNet("mvn_batch_norm_1x1", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
OCL_TEST(Test_TensorFlow, batch_norm) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, pooling) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("batch_norm", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("fused_batch_norm", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("batch_norm_text", DNN_TARGET_OPENCL, true); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("max_pool_even", targetId); |
|
|
|
|
runTensorFlowNet("max_pool_odd_valid", targetId); |
|
|
|
|
runTensorFlowNet("ave_pool_same", targetId); |
|
|
|
|
runTensorFlowNet("max_pool_odd_same", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, pooling) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, deconvolution) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("max_pool_even"); |
|
|
|
|
runTensorFlowNet("max_pool_odd_valid"); |
|
|
|
|
runTensorFlowNet("max_pool_odd_same"); |
|
|
|
|
runTensorFlowNet("ave_pool_same"); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("deconvolution", targetId); |
|
|
|
|
runTensorFlowNet("deconvolution_same", targetId); |
|
|
|
|
runTensorFlowNet("deconvolution_stride_2_same", targetId); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_valid", targetId); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_same", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, deconvolution) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, matmul) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("deconvolution"); |
|
|
|
|
runTensorFlowNet("deconvolution_same"); |
|
|
|
|
runTensorFlowNet("deconvolution_stride_2_same"); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_valid"); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_same"); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("matmul", targetId); |
|
|
|
|
runTensorFlowNet("nhwc_reshape_matmul", targetId); |
|
|
|
|
runTensorFlowNet("nhwc_transpose_reshape_matmul", targetId); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
OCL_TEST(Test_TensorFlow, deconvolution) |
|
|
|
|
TEST_P(Test_TensorFlow_layers, reshape) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("deconvolution", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("deconvolution_same", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("deconvolution_stride_2_same", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_valid", DNN_TARGET_OPENCL); |
|
|
|
|
runTensorFlowNet("deconvolution_adj_pad_same", DNN_TARGET_OPENCL); |
|
|
|
|
int targetId = GetParam(); |
|
|
|
|
runTensorFlowNet("shift_reshape_no_reorder", targetId); |
|
|
|
|
runTensorFlowNet("reshape_reduce", targetId); |
|
|
|
|
runTensorFlowNet("flatten", targetId, true); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, matmul) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("matmul"); |
|
|
|
|
runTensorFlowNet("nhwc_reshape_matmul"); |
|
|
|
|
runTensorFlowNet("nhwc_transpose_reshape_matmul"); |
|
|
|
|
} |
|
|
|
|
INSTANTIATE_TEST_CASE_P(/**/, Test_TensorFlow_layers, availableDnnTargets()); |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, defun) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("defun_dropout"); |
|
|
|
|
} |
|
|
|
|
typedef testing::TestWithParam<DNNTarget> Test_TensorFlow_nets; |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, reshape) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("shift_reshape_no_reorder"); |
|
|
|
|
runTensorFlowNet("reshape_reduce"); |
|
|
|
|
runTensorFlowNet("flatten", DNN_TARGET_CPU, true); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, fp16) |
|
|
|
|
{ |
|
|
|
|
const float l1 = 1e-3; |
|
|
|
|
const float lInf = 1e-2; |
|
|
|
|
runTensorFlowNet("fp16_single_conv", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_deconvolution", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_odd_same", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_padding_valid", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_eltwise_add_mul", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_odd_valid", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_pad_and_concat", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_even", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_padding_same", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, quantized) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("uint8_single_conv"); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, MobileNet_SSD) |
|
|
|
|
TEST_P(Test_TensorFlow_nets, MobileNet_SSD) |
|
|
|
|
{ |
|
|
|
|
std::string netPath = findDataFile("dnn/ssd_mobilenet_v1_coco.pb", false); |
|
|
|
|
std::string netConfig = findDataFile("dnn/ssd_mobilenet_v1_coco.pbtxt", false); |
|
|
|
@ -249,17 +214,20 @@ TEST(Test_TensorFlow, MobileNet_SSD) |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
Net net = readNetFromTensorflow(netPath, netConfig); |
|
|
|
|
|
|
|
|
|
net.setPreferableTarget(GetParam()); |
|
|
|
|
|
|
|
|
|
net.setInput(inp); |
|
|
|
|
|
|
|
|
|
std::vector<Mat> output; |
|
|
|
|
net.forward(output, outNames); |
|
|
|
|
|
|
|
|
|
normAssert(target[0].reshape(1, 1), output[0].reshape(1, 1)); |
|
|
|
|
normAssert(target[0].reshape(1, 1), output[0].reshape(1, 1), "", 1e-5, 1.5e-4); |
|
|
|
|
normAssert(target[1].reshape(1, 1), output[1].reshape(1, 1), "", 1e-5, 3e-4); |
|
|
|
|
normAssert(target[2].reshape(1, 1), output[2].reshape(1, 1), "", 4e-5, 1e-2); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, Inception_v2_SSD) |
|
|
|
|
TEST_P(Test_TensorFlow_nets, Inception_v2_SSD) |
|
|
|
|
{ |
|
|
|
|
std::string proto = findDataFile("dnn/ssd_inception_v2_coco_2017_11_17.pbtxt", false); |
|
|
|
|
std::string model = findDataFile("dnn/ssd_inception_v2_coco_2017_11_17.pb", false); |
|
|
|
@ -268,6 +236,8 @@ TEST(Test_TensorFlow, Inception_v2_SSD) |
|
|
|
|
Mat img = imread(findDataFile("dnn/street.png", false)); |
|
|
|
|
Mat blob = blobFromImage(img, 1.0f / 127.5, Size(300, 300), Scalar(127.5, 127.5, 127.5), true, false); |
|
|
|
|
|
|
|
|
|
net.setPreferableTarget(GetParam()); |
|
|
|
|
|
|
|
|
|
net.setInput(blob); |
|
|
|
|
// Output has shape 1x1xNx7 where N - number of detections.
|
|
|
|
|
// An every detection is a vector of values [id, classId, confidence, left, top, right, bottom]
|
|
|
|
@ -289,74 +259,57 @@ TEST(Test_TensorFlow, Inception_v2_SSD) |
|
|
|
|
normAssert(detections, ref); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
OCL_TEST(Test_TensorFlow, MobileNet_SSD) |
|
|
|
|
{ |
|
|
|
|
std::string netPath = findDataFile("dnn/ssd_mobilenet_v1_coco.pb", false); |
|
|
|
|
std::string netConfig = findDataFile("dnn/ssd_mobilenet_v1_coco.pbtxt", false); |
|
|
|
|
std::string imgPath = findDataFile("dnn/street.png", false); |
|
|
|
|
|
|
|
|
|
Mat inp; |
|
|
|
|
resize(imread(imgPath), inp, Size(300, 300)); |
|
|
|
|
inp = blobFromImage(inp, 1.0f / 127.5, Size(), Scalar(127.5, 127.5, 127.5), true); |
|
|
|
|
|
|
|
|
|
std::vector<String> outNames(3); |
|
|
|
|
outNames[0] = "concat"; |
|
|
|
|
outNames[1] = "concat_1"; |
|
|
|
|
outNames[2] = "detection_out"; |
|
|
|
|
|
|
|
|
|
std::vector<Mat> target(outNames.size()); |
|
|
|
|
for (int i = 0; i < outNames.size(); ++i) |
|
|
|
|
{ |
|
|
|
|
std::string path = findDataFile("dnn/tensorflow/ssd_mobilenet_v1_coco." + outNames[i] + ".npy", false); |
|
|
|
|
target[i] = blobFromNPY(path); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
Net net = readNetFromTensorflow(netPath, netConfig); |
|
|
|
|
|
|
|
|
|
net.setPreferableBackend(DNN_BACKEND_DEFAULT); |
|
|
|
|
net.setPreferableTarget(DNN_TARGET_OPENCL); |
|
|
|
|
|
|
|
|
|
net.setInput(inp); |
|
|
|
|
|
|
|
|
|
std::vector<Mat> output; |
|
|
|
|
net.forward(output, outNames); |
|
|
|
|
|
|
|
|
|
normAssert(target[0].reshape(1, 1), output[0].reshape(1, 1), "", 1e-5, 1.5e-4); |
|
|
|
|
normAssert(target[1].reshape(1, 1), output[1].reshape(1, 1), "", 1e-5, 3e-4); |
|
|
|
|
normAssert(target[2].reshape(1, 1), output[2].reshape(1, 1), "", 4e-5, 1e-2); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
OCL_TEST(Test_TensorFlow, Inception_v2_SSD) |
|
|
|
|
TEST_P(Test_TensorFlow_nets, opencv_face_detector_uint8) |
|
|
|
|
{ |
|
|
|
|
std::string proto = findDataFile("dnn/ssd_inception_v2_coco_2017_11_17.pbtxt", false); |
|
|
|
|
std::string model = findDataFile("dnn/ssd_inception_v2_coco_2017_11_17.pb", false); |
|
|
|
|
std::string proto = findDataFile("dnn/opencv_face_detector.pbtxt", false); |
|
|
|
|
std::string model = findDataFile("dnn/opencv_face_detector_uint8.pb", false); |
|
|
|
|
|
|
|
|
|
Net net = readNetFromTensorflow(model, proto); |
|
|
|
|
Mat img = imread(findDataFile("dnn/street.png", false)); |
|
|
|
|
Mat blob = blobFromImage(img, 1.0f / 127.5, Size(300, 300), Scalar(127.5, 127.5, 127.5), true, false); |
|
|
|
|
Mat img = imread(findDataFile("gpu/lbpcascade/er.png", false)); |
|
|
|
|
Mat blob = blobFromImage(img, 1.0, Size(), Scalar(104.0, 177.0, 123.0), false, false); |
|
|
|
|
|
|
|
|
|
net.setPreferableBackend(DNN_BACKEND_DEFAULT); |
|
|
|
|
net.setPreferableTarget(DNN_TARGET_OPENCL); |
|
|
|
|
net.setPreferableTarget(GetParam()); |
|
|
|
|
|
|
|
|
|
net.setInput(blob); |
|
|
|
|
// Output has shape 1x1xNx7 where N - number of detections.
|
|
|
|
|
// An every detection is a vector of values [id, classId, confidence, left, top, right, bottom]
|
|
|
|
|
Mat out = net.forward(); |
|
|
|
|
out = out.reshape(1, out.total() / 7); |
|
|
|
|
|
|
|
|
|
Mat detections; |
|
|
|
|
for (int i = 0; i < out.rows; ++i) |
|
|
|
|
{ |
|
|
|
|
if (out.at<float>(i, 2) > 0.5) |
|
|
|
|
detections.push_back(out.row(i).colRange(1, 7)); |
|
|
|
|
} |
|
|
|
|
// References are from test for Caffe model.
|
|
|
|
|
Mat ref = (Mat_<float>(6, 5) << 0.99520785, 0.80997437, 0.16379407, 0.87996572, 0.26685631, |
|
|
|
|
0.9934696, 0.2831718, 0.50738752, 0.345781, 0.5985168, |
|
|
|
|
0.99096733, 0.13629119, 0.24892329, 0.19756334, 0.3310290, |
|
|
|
|
0.98977017, 0.23901358, 0.09084064, 0.29902688, 0.1769477, |
|
|
|
|
0.97203469, 0.67965847, 0.06876482, 0.73999709, 0.1513494, |
|
|
|
|
0.95097077, 0.51901293, 0.45863652, 0.5777427, 0.5347801); |
|
|
|
|
normAssert(out.reshape(1, out.total() / 7).rowRange(0, 6).colRange(2, 7), ref, "", 2.8e-4, 3.4e-3); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
Mat ref = (Mat_<float>(5, 6) << 1, 0.90176028, 0.19872092, 0.36311883, 0.26461923, 0.63498729, |
|
|
|
|
3, 0.93569964, 0.64865261, 0.45906419, 0.80675775, 0.65708131, |
|
|
|
|
3, 0.75838411, 0.44668293, 0.45907149, 0.49459291, 0.52197015, |
|
|
|
|
10, 0.95932811, 0.38349164, 0.32528657, 0.40387636, 0.39165527, |
|
|
|
|
10, 0.93973452, 0.66561931, 0.37841269, 0.68074018, 0.42907384); |
|
|
|
|
normAssert(detections, ref); |
|
|
|
|
INSTANTIATE_TEST_CASE_P(/**/, Test_TensorFlow_nets, availableDnnTargets()); |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, defun) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("defun_dropout"); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, fp16) |
|
|
|
|
{ |
|
|
|
|
const float l1 = 1e-3; |
|
|
|
|
const float lInf = 1e-2; |
|
|
|
|
runTensorFlowNet("fp16_single_conv", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_deconvolution", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_odd_same", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_padding_valid", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_eltwise_add_mul", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_odd_valid", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_pad_and_concat", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_max_pool_even", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
runTensorFlowNet("fp16_padding_same", DNN_TARGET_CPU, false, l1, lInf); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, quantized) |
|
|
|
|
{ |
|
|
|
|
runTensorFlowNet("uint8_single_conv"); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, lstm) |
|
|
|
@ -390,28 +343,4 @@ TEST(Test_TensorFlow, memory_read) |
|
|
|
|
runTensorFlowNet("batch_norm_text", DNN_TARGET_CPU, true, l1, lInf, true); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
TEST(Test_TensorFlow, opencv_face_detector_uint8) |
|
|
|
|
{ |
|
|
|
|
std::string proto = findDataFile("dnn/opencv_face_detector.pbtxt", false); |
|
|
|
|
std::string model = findDataFile("dnn/opencv_face_detector_uint8.pb", false); |
|
|
|
|
|
|
|
|
|
Net net = readNetFromTensorflow(model, proto); |
|
|
|
|
Mat img = imread(findDataFile("gpu/lbpcascade/er.png", false)); |
|
|
|
|
Mat blob = blobFromImage(img, 1.0, Size(), Scalar(104.0, 177.0, 123.0), false, false); |
|
|
|
|
|
|
|
|
|
net.setInput(blob); |
|
|
|
|
// Output has shape 1x1xNx7 where N - number of detections.
|
|
|
|
|
// An every detection is a vector of values [id, classId, confidence, left, top, right, bottom]
|
|
|
|
|
Mat out = net.forward(); |
|
|
|
|
|
|
|
|
|
// References are from test for Caffe model.
|
|
|
|
|
Mat ref = (Mat_<float>(6, 5) << 0.99520785, 0.80997437, 0.16379407, 0.87996572, 0.26685631, |
|
|
|
|
0.9934696, 0.2831718, 0.50738752, 0.345781, 0.5985168, |
|
|
|
|
0.99096733, 0.13629119, 0.24892329, 0.19756334, 0.3310290, |
|
|
|
|
0.98977017, 0.23901358, 0.09084064, 0.29902688, 0.1769477, |
|
|
|
|
0.97203469, 0.67965847, 0.06876482, 0.73999709, 0.1513494, |
|
|
|
|
0.95097077, 0.51901293, 0.45863652, 0.5777427, 0.5347801); |
|
|
|
|
normAssert(out.reshape(1, out.total() / 7).rowRange(0, 6).colRange(2, 7), ref, "", 2.8e-4, 3.4e-3); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
} |
|
|
|
|