|
|
|
# Ultralytics YOLO 🚀, AGPL-3.0 license
|
|
|
|
|
|
|
|
import pytest
|
|
|
|
import torch
|
|
|
|
|
|
|
|
from ultralytics import YOLO
|
|
|
|
from ultralytics.utils import ASSETS, WEIGHTS_DIR, checks
|
|
|
|
|
|
|
|
CUDA_IS_AVAILABLE = checks.cuda_is_available()
|
|
|
|
CUDA_DEVICE_COUNT = checks.cuda_device_count()
|
|
|
|
|
|
|
|
MODEL = WEIGHTS_DIR / "path with spaces" / "yolov8n.pt" # test spaces in path
|
|
|
|
DATA = "coco8.yaml"
|
|
|
|
BUS = ASSETS / "bus.jpg"
|
|
|
|
|
|
|
|
|
|
|
|
def test_checks():
|
|
|
|
"""Validate CUDA settings against torch CUDA functions."""
|
|
|
|
assert torch.cuda.is_available() == CUDA_IS_AVAILABLE
|
|
|
|
assert torch.cuda.device_count() == CUDA_DEVICE_COUNT
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.skipif(not CUDA_IS_AVAILABLE, reason="CUDA is not available")
|
|
|
|
def test_train():
|
|
|
|
"""Test model training on a minimal dataset."""
|
|
|
|
device = 0 if CUDA_DEVICE_COUNT == 1 else [0, 1]
|
|
|
|
YOLO(MODEL).train(data=DATA, imgsz=64, epochs=1, device=device) # requires imgsz>=64
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.slow
|
|
|
|
@pytest.mark.skipif(not CUDA_IS_AVAILABLE, reason="CUDA is not available")
|
|
|
|
def test_predict_multiple_devices():
|
|
|
|
"""Validate model prediction on multiple devices."""
|
|
|
|
model = YOLO("yolov8n.pt")
|
|
|
|
model = model.cpu()
|
|
|
|
assert str(model.device) == "cpu"
|
|
|
|
_ = model(BUS) # CPU inference
|
|
|
|
assert str(model.device) == "cpu"
|
|
|
|
|
|
|
|
model = model.to("cuda:0")
|
|
|
|
assert str(model.device) == "cuda:0"
|
|
|
|
_ = model(BUS) # CUDA inference
|
|
|
|
assert str(model.device) == "cuda:0"
|
|
|
|
|
|
|
|
model = model.cpu()
|
|
|
|
assert str(model.device) == "cpu"
|
|
|
|
_ = model(BUS) # CPU inference
|
|
|
|
assert str(model.device) == "cpu"
|
|
|
|
|
|
|
|
model = model.cuda()
|
|
|
|
assert str(model.device) == "cuda:0"
|
|
|
|
_ = model(BUS) # CUDA inference
|
|
|
|
assert str(model.device) == "cuda:0"
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.skipif(not CUDA_IS_AVAILABLE, reason="CUDA is not available")
|
|
|
|
def test_autobatch():
|
|
|
|
"""Check batch size for YOLO model using autobatch."""
|
|
|
|
from ultralytics.utils.autobatch import check_train_batch_size
|
|
|
|
|
|
|
|
check_train_batch_size(YOLO(MODEL).model.cuda(), imgsz=128, amp=True)
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.slow
|
|
|
|
@pytest.mark.skipif(not CUDA_IS_AVAILABLE, reason="CUDA is not available")
|
|
|
|
def test_utils_benchmarks():
|
|
|
|
"""Profile YOLO models for performance benchmarks."""
|
|
|
|
from ultralytics.utils.benchmarks import ProfileModels
|
|
|
|
|
|
|
|
# Pre-export a dynamic engine model to use dynamic inference
|
|
|
|
YOLO(MODEL).export(format="engine", imgsz=32, dynamic=True, batch=1)
|
|
|
|
ProfileModels([MODEL], imgsz=32, half=False, min_time=1, num_timed_runs=3, num_warmup_runs=1).profile()
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.skipif(not CUDA_IS_AVAILABLE, reason="CUDA is not available")
|
|
|
|
def test_predict_sam():
|
|
|
|
"""Test SAM model prediction with various prompts."""
|
|
|
|
from ultralytics import SAM
|
|
|
|
from ultralytics.models.sam import Predictor as SAMPredictor
|
|
|
|
|
|
|
|
# Load a model
|
|
|
|
model = SAM(WEIGHTS_DIR / "sam_b.pt")
|
|
|
|
|
|
|
|
# Display model information (optional)
|
|
|
|
model.info()
|
|
|
|
|
|
|
|
# Run inference
|
|
|
|
model(BUS, device=0)
|
|
|
|
|
|
|
|
# Run inference with bboxes prompt
|
|
|
|
model(BUS, bboxes=[439, 437, 524, 709], device=0)
|
|
|
|
|
|
|
|
# Run inference with points prompt
|
|
|
|
model(ASSETS / "zidane.jpg", points=[900, 370], labels=[1], device=0)
|
|
|
|
|
|
|
|
# Create SAMPredictor
|
|
|
|
overrides = dict(conf=0.25, task="segment", mode="predict", imgsz=1024, model=WEIGHTS_DIR / "mobile_sam.pt")
|
|
|
|
predictor = SAMPredictor(overrides=overrides)
|
|
|
|
|
|
|
|
# Set image
|
|
|
|
predictor.set_image(ASSETS / "zidane.jpg") # set with image file
|
|
|
|
# predictor(bboxes=[439, 437, 524, 709])
|
|
|
|
# predictor(points=[900, 370], labels=[1])
|
|
|
|
|
|
|
|
# Reset image
|
|
|
|
predictor.reset_image()
|