You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 

8.6 KiB

comments description
true Discover how to integrate hyperparameter tuning with Ray Tune and Ultralytics YOLOv8. Speed up the tuning process and optimize your model's performance.

Hyperparameter Tuning with Ray Tune and YOLOv8

Hyperparameter tuning (or hyperparameter optimization) is the process of determining the right combination of hyperparameters that maximizes model performance. It works by running multiple trials in a single training process, evaluating the performance of each trial, and selecting the best hyperparameter values based on the evaluation results.

Ultralytics YOLOv8 and Ray Tune Integration

Ultralytics YOLOv8 integrates hyperparameter tuning with Ray Tune, allowing you to easily optimize your YOLOv8 model's hyperparameters. By using Ray Tune, you can leverage advanced search algorithms, parallelism, and early stopping to speed up the tuning process and achieve better model performance.

Ray Tune

Ray Tune is a powerful and flexible hyperparameter tuning library for machine learning models. It provides an efficient way to optimize hyperparameters by supporting various search algorithms, parallelism, and early stopping strategies. Ray Tune's flexible architecture enables seamless integration with popular machine learning frameworks, including Ultralytics YOLOv8.

Weights & Biases

YOLOv8 also supports optional integration with Weights & Biases (wandb) for tracking the tuning progress.

Installation

To install the required packages, run:

!!! tip "Installation"

```bash
pip install -U ultralytics "ray[tune]"  # install and/or update
pip install wandb  # optional
```

Usage

!!! example "Usage"

```python
from ultralytics import YOLO

model = YOLO("yolov8n.pt")
results = model.tune(data="coco128.yaml")
```

tune() Method Parameters

The tune() method in YOLOv8 provides an easy-to-use interface for hyperparameter tuning with Ray Tune. It accepts several arguments that allow you to customize the tuning process. Below is a detailed explanation of each parameter:

Parameter Type Description Default Value
data str The dataset configuration file (in YAML format) to run the tuner on. This file should specify the training and validation data paths, as well as other dataset-specific settings.
space dict, optional A dictionary defining the hyperparameter search space for Ray Tune. Each key corresponds to a hyperparameter name, and the value specifies the range of values to explore during tuning. If not provided, YOLOv8 uses a default search space with various hyperparameters.
grace_period int, optional The grace period in epochs for the ASHA scheduler in Ray Tune. The scheduler will not terminate any trial before this number of epochs, allowing the model to have some minimum training before making a decision on early stopping. 10
gpu_per_trial int, optional The number of GPUs to allocate per trial during tuning. This helps manage GPU usage, particularly in multi-GPU environments. If not provided, the tuner will use all available GPUs. None
max_samples int, optional The maximum number of trials to run during tuning. This parameter helps control the total number of hyperparameter combinations tested, ensuring the tuning process does not run indefinitely. 10
train_args dict, optional A dictionary of additional arguments to pass to the train() method during tuning. These arguments can include settings like the number of training epochs, batch size, and other training-specific configurations. {}

By customizing these parameters, you can fine-tune the hyperparameter optimization process to suit your specific needs and available computational resources.

Default Search Space Description

The following table lists the default search space parameters for hyperparameter tuning in YOLOv8 with Ray Tune. Each parameter has a specific value range defined by tune.uniform().

Parameter Value Range Description
lr0 tune.uniform(1e-5, 1e-1) Initial learning rate
lrf tune.uniform(0.01, 1.0) Final learning rate factor
momentum tune.uniform(0.6, 0.98) Momentum
weight_decay tune.uniform(0.0, 0.001) Weight decay
warmup_epochs tune.uniform(0.0, 5.0) Warmup epochs
warmup_momentum tune.uniform(0.0, 0.95) Warmup momentum
box tune.uniform(0.02, 0.2) Box loss weight
cls tune.uniform(0.2, 4.0) Class loss weight
hsv_h tune.uniform(0.0, 0.1) Hue augmentation range
hsv_s tune.uniform(0.0, 0.9) Saturation augmentation range
hsv_v tune.uniform(0.0, 0.9) Value (brightness) augmentation range
degrees tune.uniform(0.0, 45.0) Rotation augmentation range (degrees)
translate tune.uniform(0.0, 0.9) Translation augmentation range
scale tune.uniform(0.0, 0.9) Scaling augmentation range
shear tune.uniform(0.0, 10.0) Shear augmentation range (degrees)
perspective tune.uniform(0.0, 0.001) Perspective augmentation range
flipud tune.uniform(0.0, 1.0) Vertical flip augmentation probability
fliplr tune.uniform(0.0, 1.0) Horizontal flip augmentation probability
mosaic tune.uniform(0.0, 1.0) Mosaic augmentation probability
mixup tune.uniform(0.0, 1.0) Mixup augmentation probability
copy_paste tune.uniform(0.0, 1.0) Copy-paste augmentation probability

Custom Search Space Example

In this example, we demonstrate how to use a custom search space for hyperparameter tuning with Ray Tune and YOLOv8. By providing a custom search space, you can focus the tuning process on specific hyperparameters of interest.

!!! example "Usage"

```python
from ultralytics import YOLO
from ray import tune

model = YOLO("yolov8n.pt")
result = model.tune(
    data="coco128.yaml",
    space={"lr0": tune.uniform(1e-5, 1e-1)},
    train_args={"epochs": 50}
)
```

In the code snippet above, we create a YOLO model with the "yolov8n.pt" pretrained weights. Then, we call the tune() method, specifying the dataset configuration with "coco128.yaml". We provide a custom search space for the initial learning rate lr0 using a dictionary with the key "lr0" and the value tune.uniform(1e-5, 1e-1). Finally, we pass additional training arguments, such as the number of epochs, using the train_args parameter.