diff --git a/docs/en/datasets/classify/caltech101.md b/docs/en/datasets/classify/caltech101.md
index 2462a167fc..b50a51916b 100644
--- a/docs/en/datasets/classify/caltech101.md
+++ b/docs/en/datasets/classify/caltech101.md
@@ -28,7 +28,7 @@ The Caltech-101 dataset is extensively used for training and evaluating deep lea
To train a YOLO model on the Caltech-101 dataset for 100 epochs, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -61,7 +61,7 @@ The example showcases the variety and complexity of the objects in the Caltech-1
If you use the Caltech-101 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -90,7 +90,7 @@ The [Caltech-101](https://data.caltech.edu/records/mzrjq-6wc02) dataset is widel
To train an Ultralytics YOLO model on the Caltech-101 dataset, you can use the provided code snippets. For example, to train for 100 epochs:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -128,7 +128,7 @@ These features make it an excellent choice for training and evaluating object re
Citing the Caltech-101 dataset in your research acknowledges the creators' contributions and provides a reference for others who might use the dataset. The recommended citation is:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/classify/caltech256.md b/docs/en/datasets/classify/caltech256.md
index a2551b9a60..c337721009 100644
--- a/docs/en/datasets/classify/caltech256.md
+++ b/docs/en/datasets/classify/caltech256.md
@@ -39,7 +39,7 @@ The Caltech-256 dataset is extensively used for training and evaluating deep lea
To train a YOLO model on the Caltech-256 dataset for 100 epochs, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -72,7 +72,7 @@ The example showcases the diversity and complexity of the objects in the Caltech
If you use the Caltech-256 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -98,7 +98,7 @@ The [Caltech-256](https://data.caltech.edu/records/nyy15-4j048) dataset is a lar
To train a YOLO model on the Caltech-256 dataset for 100 epochs, you can use the following code snippets. Refer to the model [Training](../../modes/train.md) page for additional options.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/classify/cifar10.md b/docs/en/datasets/classify/cifar10.md
index 39762681b2..865c80865f 100644
--- a/docs/en/datasets/classify/cifar10.md
+++ b/docs/en/datasets/classify/cifar10.md
@@ -42,7 +42,7 @@ The CIFAR-10 dataset is widely used for training and evaluating deep learning mo
To train a YOLO model on the CIFAR-10 dataset for 100 epochs with an image size of 32x32, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -75,7 +75,7 @@ The example showcases the variety and complexity of the objects in the CIFAR-10
If you use the CIFAR-10 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -96,7 +96,7 @@ We would like to acknowledge Alex Krizhevsky for creating and maintaining the CI
To train a YOLO model on the CIFAR-10 dataset using Ultralytics, you can follow the examples provided for both Python and CLI. Here is a basic example to train your model for 100 epochs with an image size of 32x32 pixels:
-!!! Example
+!!! example
=== "Python"
@@ -153,7 +153,7 @@ Each subset comprises images categorized into 10 classes, with their annotations
If you use the CIFAR-10 dataset in your research or development projects, make sure to cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/classify/cifar100.md b/docs/en/datasets/classify/cifar100.md
index 2861c9469f..ca868240f7 100644
--- a/docs/en/datasets/classify/cifar100.md
+++ b/docs/en/datasets/classify/cifar100.md
@@ -31,7 +31,7 @@ The CIFAR-100 dataset is extensively used for training and evaluating deep learn
To train a YOLO model on the CIFAR-100 dataset for 100 epochs with an image size of 32x32, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -64,7 +64,7 @@ The example showcases the variety and complexity of the objects in the CIFAR-100
If you use the CIFAR-100 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -89,7 +89,7 @@ The [CIFAR-100 dataset](https://www.cs.toronto.edu/~kriz/cifar.html) is a large
You can train a YOLO model on the CIFAR-100 dataset using either Python or CLI commands. Here's how:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/classify/fashion-mnist.md b/docs/en/datasets/classify/fashion-mnist.md
index 2de2a805c3..d1c87e2676 100644
--- a/docs/en/datasets/classify/fashion-mnist.md
+++ b/docs/en/datasets/classify/fashion-mnist.md
@@ -56,7 +56,7 @@ The Fashion-MNIST dataset is widely used for training and evaluating deep learni
To train a CNN model on the Fashion-MNIST dataset for 100 epochs with an image size of 28x28, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -99,7 +99,7 @@ The [Fashion-MNIST](https://github.com/zalandoresearch/fashion-mnist) dataset is
To train an Ultralytics YOLO model on the Fashion-MNIST dataset, you can use both Python and CLI commands. Here's a quick example to get you started:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/classify/imagenet.md b/docs/en/datasets/classify/imagenet.md
index ae1ade9ba3..8c4102caf8 100644
--- a/docs/en/datasets/classify/imagenet.md
+++ b/docs/en/datasets/classify/imagenet.md
@@ -41,7 +41,7 @@ The ImageNet dataset is widely used for training and evaluating deep learning mo
To train a deep learning model on the ImageNet dataset for 100 epochs with an image size of 224x224, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -74,7 +74,7 @@ The example showcases the variety and complexity of the images in the ImageNet d
If you use the ImageNet dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -102,7 +102,7 @@ The [ImageNet dataset](https://www.image-net.org/) is a large-scale database con
To use a pretrained Ultralytics YOLO model for image classification on the ImageNet dataset, follow these steps:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/classify/imagenet10.md b/docs/en/datasets/classify/imagenet10.md
index 38764c89ec..b74e2a7a62 100644
--- a/docs/en/datasets/classify/imagenet10.md
+++ b/docs/en/datasets/classify/imagenet10.md
@@ -27,7 +27,7 @@ The ImageNet10 dataset is useful for quickly testing and debugging computer visi
To test a deep learning model on the ImageNet10 dataset with an image size of 224x224, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Test Example"
+!!! example "Test Example"
=== "Python"
@@ -58,7 +58,7 @@ The ImageNet10 dataset contains a subset of images from the original ImageNet da
If you use the ImageNet10 dataset in your research or development work, please cite the original ImageNet paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -86,7 +86,7 @@ The [ImageNet10](https://github.com/ultralytics/assets/releases/download/v0.0.0/
To test your deep learning model on the ImageNet10 dataset with an image size of 224x224, use the following code snippets.
-!!! Example "Test Example"
+!!! example "Test Example"
=== "Python"
diff --git a/docs/en/datasets/classify/imagenette.md b/docs/en/datasets/classify/imagenette.md
index fa06e0d38f..42368a672a 100644
--- a/docs/en/datasets/classify/imagenette.md
+++ b/docs/en/datasets/classify/imagenette.md
@@ -29,7 +29,7 @@ The ImageNette dataset is widely used for training and evaluating deep learning
To train a model on the ImageNette dataset for 100 epochs with a standard image size of 224x224, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -64,7 +64,7 @@ For faster prototyping and training, the ImageNette dataset is also available in
To use these datasets, simply replace 'imagenette' with 'imagenette160' or 'imagenette320' in the training command. The following code snippets illustrate this:
-!!! Example "Train Example with ImageNette160"
+!!! example "Train Example with ImageNette160"
=== "Python"
@@ -85,7 +85,7 @@ To use these datasets, simply replace 'imagenette' with 'imagenette160' or 'imag
yolo classify train data=imagenette160 model=yolov8n-cls.pt epochs=100 imgsz=160
```
-!!! Example "Train Example with ImageNette320"
+!!! example "Train Example with ImageNette320"
=== "Python"
@@ -122,7 +122,7 @@ The [ImageNette dataset](https://github.com/fastai/imagenette) is a simplified s
To train a YOLO model on the ImageNette dataset for 100 epochs, you can use the following commands. Make sure to have the Ultralytics YOLO environment set up.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -159,7 +159,7 @@ For more details on model training and dataset management, explore the [Dataset
Yes, the ImageNette dataset is also available in two resized versions: ImageNette160 and ImageNette320. These versions help in faster prototyping and are especially useful when computational resources are limited.
-!!! Example "Train Example with ImageNette160"
+!!! example "Train Example with ImageNette160"
=== "Python"
diff --git a/docs/en/datasets/classify/imagewoof.md b/docs/en/datasets/classify/imagewoof.md
index 0f6537453b..3abb512a0d 100644
--- a/docs/en/datasets/classify/imagewoof.md
+++ b/docs/en/datasets/classify/imagewoof.md
@@ -26,7 +26,7 @@ The ImageWoof dataset is widely used for training and evaluating deep learning m
To train a CNN model on the ImageWoof dataset for 100 epochs with an image size of 224x224, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -59,7 +59,7 @@ ImageWoof dataset comes in three different sizes to accommodate various research
To use these variants in your training, simply replace 'imagewoof' in the dataset argument with 'imagewoof320' or 'imagewoof160'. For example:
-!!! Example "Example"
+!!! example "Example"
=== "Python"
@@ -109,7 +109,7 @@ The [ImageWoof](https://github.com/fastai/imagenette) dataset is a challenging s
To train a Convolutional Neural Network (CNN) model on the ImageWoof dataset using Ultralytics YOLO for 100 epochs at an image size of 224x224, you can use the following code:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/classify/index.md b/docs/en/datasets/classify/index.md
index bc3d719124..2b01824a96 100644
--- a/docs/en/datasets/classify/index.md
+++ b/docs/en/datasets/classify/index.md
@@ -78,7 +78,7 @@ This structured approach ensures that the model can effectively learn from well-
## Usage
-!!! Example
+!!! example
=== "Python"
@@ -194,7 +194,7 @@ For additional insights and real-world applications, you can explore [Ultralytic
Training a model using Ultralytics YOLO can be done easily in both Python and CLI. Here's an example:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/classify/mnist.md b/docs/en/datasets/classify/mnist.md
index 6fcf5bd436..7ee4531392 100644
--- a/docs/en/datasets/classify/mnist.md
+++ b/docs/en/datasets/classify/mnist.md
@@ -34,7 +34,7 @@ The MNIST dataset is widely used for training and evaluating deep learning model
To train a CNN model on the MNIST dataset for 100 epochs with an image size of 32x32, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -69,7 +69,7 @@ If you use the MNIST dataset in your
research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -95,7 +95,7 @@ The [MNIST](http://yann.lecun.com/exdb/mnist/) dataset, or Modified National Ins
To train a model on the MNIST dataset using Ultralytics YOLO, you can follow these steps:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/african-wildlife.md b/docs/en/datasets/detect/african-wildlife.md
index fce0cf54f3..d0c86a0314 100644
--- a/docs/en/datasets/detect/african-wildlife.md
+++ b/docs/en/datasets/detect/african-wildlife.md
@@ -35,7 +35,7 @@ This dataset can be applied in various computer vision tasks such as object dete
A YAML (Yet Another Markup Language) file defines the dataset configuration, including paths, classes, and other pertinent details. For the African wildlife dataset, the `african-wildlife.yaml` file is located at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/african-wildlife.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/african-wildlife.yaml).
-!!! Example "ultralytics/cfg/datasets/african-wildlife.yaml"
+!!! example "ultralytics/cfg/datasets/african-wildlife.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/african-wildlife.yaml"
@@ -45,7 +45,7 @@ A YAML (Yet Another Markup Language) file defines the dataset configuration, inc
To train a YOLOv8n model on the African wildlife dataset for 100 epochs with an image size of 640, use the provided code samples. For a comprehensive list of available parameters, refer to the model's [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -66,7 +66,7 @@ To train a YOLOv8n model on the African wildlife dataset for 100 epochs with an
yolo detect train data=african-wildlife.yaml model=yolov8n.pt epochs=100 imgsz=640
```
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
@@ -111,7 +111,7 @@ The African Wildlife Dataset includes images of four common animal species found
You can train a YOLOv8 model on the African Wildlife Dataset by using the `african-wildlife.yaml` configuration file. Below is an example of how to train the YOLOv8n model for 100 epochs with an image size of 640:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/detect/argoverse.md b/docs/en/datasets/detect/argoverse.md
index c3a2c6e232..47ef822b08 100644
--- a/docs/en/datasets/detect/argoverse.md
+++ b/docs/en/datasets/detect/argoverse.md
@@ -8,7 +8,7 @@ keywords: Argoverse dataset, autonomous driving, 3D tracking, motion forecasting
The [Argoverse](https://www.argoverse.org/) dataset is a collection of data designed to support research in autonomous driving tasks, such as 3D tracking, motion forecasting, and stereo depth estimation. Developed by Argo AI, the dataset provides a wide range of high-quality sensor data, including high-resolution images, LiDAR point clouds, and map data.
-!!! Note
+!!! note
The Argoverse dataset `*.zip` file required for training was removed from Amazon S3 after the shutdown of Argo AI by Ford, but we have made it available for manual download on [Google Drive](https://drive.google.com/file/d/1st9qW3BeIwQsnR0t8mRpvbsSWIo16ACi/view?usp=drive_link).
@@ -35,7 +35,7 @@ The Argoverse dataset is widely used for training and evaluating deep learning m
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. For the case of the Argoverse dataset, the `Argoverse.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/Argoverse.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/Argoverse.yaml).
-!!! Example "ultralytics/cfg/datasets/Argoverse.yaml"
+!!! example "ultralytics/cfg/datasets/Argoverse.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/Argoverse.yaml"
@@ -45,7 +45,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the Argoverse dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -80,7 +80,7 @@ The example showcases the variety and complexity of the data in the Argoverse da
If you use the Argoverse dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -106,7 +106,7 @@ The [Argoverse](https://www.argoverse.org/) dataset, developed by Argo AI, suppo
To train a YOLOv8 model with the Argoverse dataset, use the provided YAML configuration file and the following code:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/brain-tumor.md b/docs/en/datasets/detect/brain-tumor.md
index 38d69ff61a..aa48c5f9e4 100644
--- a/docs/en/datasets/detect/brain-tumor.md
+++ b/docs/en/datasets/detect/brain-tumor.md
@@ -34,7 +34,7 @@ The application of brain tumor detection using computer vision enables early dia
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the brain tumor dataset, the `brain-tumor.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/brain-tumor.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/brain-tumor.yaml).
-!!! Example "ultralytics/cfg/datasets/brain-tumor.yaml"
+!!! example "ultralytics/cfg/datasets/brain-tumor.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/brain-tumor.yaml"
@@ -44,7 +44,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the brain tumor dataset for 100 epochs with an image size of 640, utilize the provided code snippets. For a detailed list of available arguments, consult the model's [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -65,7 +65,7 @@ To train a YOLOv8n model on the brain tumor dataset for 100 epochs with an image
yolo detect train data=brain-tumor.yaml model=yolov8n.pt epochs=100 imgsz=640
```
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
@@ -110,7 +110,7 @@ The brain tumor dataset is divided into two subsets: the **training set** consis
You can train a YOLOv8 model on the brain tumor dataset for 100 epochs with an image size of 640px using both Python and CLI methods. Below are the examples for both:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -142,7 +142,7 @@ Using the brain tumor dataset in AI projects enables early diagnosis and treatme
Inference using a fine-tuned YOLOv8 model can be performed with either Python or CLI approaches. Here are the examples:
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
diff --git a/docs/en/datasets/detect/coco.md b/docs/en/datasets/detect/coco.md
index b0d42b9bfb..25878ed913 100644
--- a/docs/en/datasets/detect/coco.md
+++ b/docs/en/datasets/detect/coco.md
@@ -52,7 +52,7 @@ The COCO dataset is widely used for training and evaluating deep learning models
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO dataset, the `coco.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml).
-!!! Example "ultralytics/cfg/datasets/coco.yaml"
+!!! example "ultralytics/cfg/datasets/coco.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco.yaml"
@@ -62,7 +62,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the COCO dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -97,7 +97,7 @@ The example showcases the variety and complexity of the images in the COCO datas
If you use the COCO dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -124,7 +124,7 @@ The [COCO dataset](https://cocodataset.org/#home) (Common Objects in Context) is
To train a YOLOv8 model using the COCO dataset, you can use the following code snippets:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/coco8.md b/docs/en/datasets/detect/coco8.md
index b2ed77b4e5..a0972693ea 100644
--- a/docs/en/datasets/detect/coco8.md
+++ b/docs/en/datasets/detect/coco8.md
@@ -27,7 +27,7 @@ This dataset is intended for use with Ultralytics [HUB](https://hub.ultralytics.
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO8 dataset, the `coco8.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8.yaml).
-!!! Example "ultralytics/cfg/datasets/coco8.yaml"
+!!! example "ultralytics/cfg/datasets/coco8.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco8.yaml"
@@ -37,7 +37,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the COCO8 dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -72,7 +72,7 @@ The example showcases the variety and complexity of the images in the COCO8 data
If you use the COCO dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -99,7 +99,7 @@ The Ultralytics COCO8 dataset is a compact yet versatile object detection datase
To train a YOLOv8 model using the COCO8 dataset, you can employ either Python or CLI commands. Here's how you can start:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/globalwheat2020.md b/docs/en/datasets/detect/globalwheat2020.md
index 8b8a0467ad..37b9759f36 100644
--- a/docs/en/datasets/detect/globalwheat2020.md
+++ b/docs/en/datasets/detect/globalwheat2020.md
@@ -30,7 +30,7 @@ The Global Wheat Head Dataset is widely used for training and evaluating deep le
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. For the case of the Global Wheat Head Dataset, the `GlobalWheat2020.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/GlobalWheat2020.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/GlobalWheat2020.yaml).
-!!! Example "ultralytics/cfg/datasets/GlobalWheat2020.yaml"
+!!! example "ultralytics/cfg/datasets/GlobalWheat2020.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/GlobalWheat2020.yaml"
@@ -40,7 +40,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the Global Wheat Head Dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -75,7 +75,7 @@ The example showcases the variety and complexity of the data in the Global Wheat
If you use the Global Wheat Head Dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -100,7 +100,7 @@ The Global Wheat Head Dataset is primarily used for developing and training deep
To train a YOLOv8n model on the Global Wheat Head Dataset, you can use the following code snippets. Make sure you have the `GlobalWheat2020.yaml` configuration file specifying dataset paths and classes:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/index.md b/docs/en/datasets/detect/index.md
index e43cd46104..f76ba40bc2 100644
--- a/docs/en/datasets/detect/index.md
+++ b/docs/en/datasets/detect/index.md
@@ -48,7 +48,7 @@ When using the Ultralytics YOLO format, organize your training and validation im
Here's how you can use these formats to train your model:
-!!! Example
+!!! example
=== "Python"
@@ -100,7 +100,7 @@ If you have your own dataset and would like to use it for training detection mod
You can easily convert labels from the popular COCO dataset format to the YOLO format using the following code snippet:
-!!! Example
+!!! example
=== "Python"
@@ -164,7 +164,7 @@ Each dataset page provides detailed information on the structure and usage tailo
To start training a YOLOv8 model, ensure your dataset is formatted correctly and the paths are defined in a YAML file. Use the following script to begin training:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/detect/lvis.md b/docs/en/datasets/detect/lvis.md
index afc91fc680..8c06920fd6 100644
--- a/docs/en/datasets/detect/lvis.md
+++ b/docs/en/datasets/detect/lvis.md
@@ -48,7 +48,7 @@ The LVIS dataset is widely used for training and evaluating deep learning models
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the LVIS dataset, the `lvis.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/lvis.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/lvis.yaml).
-!!! Example "ultralytics/cfg/datasets/lvis.yaml"
+!!! example "ultralytics/cfg/datasets/lvis.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/lvis.yaml"
@@ -58,7 +58,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the LVIS dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -93,7 +93,7 @@ The example showcases the variety and complexity of the images in the LVIS datas
If you use the LVIS dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -118,7 +118,7 @@ The [LVIS dataset](https://www.lvisdataset.org/) is a large-scale dataset with f
To train a YOLOv8n model on the LVIS dataset for 100 epochs with an image size of 640, follow the example below. This process utilizes Ultralytics' framework, which offers comprehensive training features.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/objects365.md b/docs/en/datasets/detect/objects365.md
index ea95798fe1..f2b44f245a 100644
--- a/docs/en/datasets/detect/objects365.md
+++ b/docs/en/datasets/detect/objects365.md
@@ -30,7 +30,7 @@ The Objects365 dataset is widely used for training and evaluating deep learning
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. For the case of the Objects365 Dataset, the `Objects365.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/Objects365.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/Objects365.yaml).
-!!! Example "ultralytics/cfg/datasets/Objects365.yaml"
+!!! example "ultralytics/cfg/datasets/Objects365.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/Objects365.yaml"
@@ -40,7 +40,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the Objects365 dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -75,7 +75,7 @@ The example showcases the variety and complexity of the data in the Objects365 d
If you use the Objects365 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -101,7 +101,7 @@ The [Objects365 dataset](https://www.objects365.org/) is designed for object det
To train a YOLOv8n model using the Objects365 dataset for 100 epochs with an image size of 640, follow these instructions:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/open-images-v7.md b/docs/en/datasets/detect/open-images-v7.md
index 92958773e9..f6b0c63bc4 100644
--- a/docs/en/datasets/detect/open-images-v7.md
+++ b/docs/en/datasets/detect/open-images-v7.md
@@ -61,7 +61,7 @@ Open Images V7 is a cornerstone for training and evaluating state-of-the-art mod
Typically, datasets come with a YAML (Yet Another Markup Language) file that delineates the dataset's configuration. For the case of Open Images V7, a hypothetical `OpenImagesV7.yaml` might exist. For accurate paths and configurations, one should refer to the dataset's official repository or documentation.
-!!! Example "OpenImagesV7.yaml"
+!!! example "OpenImagesV7.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/open-images-v7.yaml"
@@ -71,7 +71,7 @@ Typically, datasets come with a YAML (Yet Another Markup Language) file that del
To train a YOLOv8n model on the Open Images V7 dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Warning
+!!! warning
The complete Open Images V7 dataset comprises 1,743,042 training images and 41,620 validation images, requiring approximately **561 GB of storage space** upon download.
@@ -80,7 +80,7 @@ To train a YOLOv8n model on the Open Images V7 dataset for 100 epochs with an im
- Verify that your device has enough storage capacity.
- Ensure a robust and speedy internet connection.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -115,7 +115,7 @@ Researchers can gain invaluable insights into the array of computer vision chall
For those employing Open Images V7 in their work, it's prudent to cite the relevant papers and acknowledge the creators:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -140,7 +140,7 @@ Open Images V7 is an extensive and versatile dataset created by Google, designed
To train a YOLOv8 model on the Open Images V7 dataset, you can use both Python and CLI commands. Here's an example of training the YOLOv8n model for 100 epochs with an image size of 640:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/roboflow-100.md b/docs/en/datasets/detect/roboflow-100.md
index 870ecb842a..844326c381 100644
--- a/docs/en/datasets/detect/roboflow-100.md
+++ b/docs/en/datasets/detect/roboflow-100.md
@@ -37,11 +37,11 @@ This structure enables a diverse and extensive testing ground for object detecti
Dataset benchmarking evaluates machine learning model performance on specific datasets using standardized metrics like accuracy, mean average precision and F1-score.
-!!! Tip "Benchmarking"
+!!! tip "Benchmarking"
Benchmarking results will be stored in "ultralytics-benchmarks/evaluation.txt"
-!!! Example "Benchmarking example"
+!!! example "Benchmarking example"
=== "Python"
@@ -113,7 +113,7 @@ The diversity in the Roboflow 100 benchmark that can be seen above is a signific
If you use the Roboflow 100 dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -139,7 +139,7 @@ The **Roboflow 100** dataset, developed by [Roboflow](https://roboflow.com/?ref=
To use the Roboflow 100 dataset for benchmarking, you can implement the RF100Benchmark class from the Ultralytics library. Here's a brief example:
-!!! Example "Benchmarking example"
+!!! example "Benchmarking example"
=== "Python"
@@ -203,7 +203,7 @@ The **Roboflow 100** dataset is accessible on [GitHub](https://github.com/robofl
When using the Roboflow 100 dataset in your research, ensure to properly cite it. Here is the recommended citation:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/detect/signature.md b/docs/en/datasets/detect/signature.md
index db82a59712..0d76e11f50 100644
--- a/docs/en/datasets/detect/signature.md
+++ b/docs/en/datasets/detect/signature.md
@@ -23,7 +23,7 @@ This dataset can be applied in various computer vision tasks such as object dete
A YAML (Yet Another Markup Language) file defines the dataset configuration, including paths and classes information. For the signature detection dataset, the `signature.yaml` file is located at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/signature.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/signature.yaml).
-!!! Example "ultralytics/cfg/datasets/signature.yaml"
+!!! example "ultralytics/cfg/datasets/signature.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/signature.yaml"
@@ -33,7 +33,7 @@ A YAML (Yet Another Markup Language) file defines the dataset configuration, inc
To train a YOLOv8n model on the signature detection dataset for 100 epochs with an image size of 640, use the provided code samples. For a comprehensive list of available parameters, refer to the model's [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -54,7 +54,7 @@ To train a YOLOv8n model on the signature detection dataset for 100 epochs with
yolo detect train data=signature.yaml model=yolov8n.pt epochs=100 imgsz=640
```
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
@@ -102,7 +102,7 @@ To train a YOLOv8n model on the Signature Detection Dataset, follow these steps:
1. Download the `signature.yaml` dataset configuration file from [signature.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/signature.yaml).
2. Use the following Python script or CLI command to start training:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -140,7 +140,7 @@ To perform inference using a model trained on the Signature Detection Dataset, f
1. Load your fine-tuned model.
2. Use the below Python script or CLI command to perform inference:
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
diff --git a/docs/en/datasets/detect/sku-110k.md b/docs/en/datasets/detect/sku-110k.md
index b307e5974e..145468321e 100644
--- a/docs/en/datasets/detect/sku-110k.md
+++ b/docs/en/datasets/detect/sku-110k.md
@@ -43,7 +43,7 @@ The SKU-110k dataset is widely used for training and evaluating deep learning mo
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. For the case of the SKU-110K dataset, the `SKU-110K.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/SKU-110K.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/SKU-110K.yaml).
-!!! Example "ultralytics/cfg/datasets/SKU-110K.yaml"
+!!! example "ultralytics/cfg/datasets/SKU-110K.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/SKU-110K.yaml"
@@ -53,7 +53,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the SKU-110K dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -88,7 +88,7 @@ The example showcases the variety and complexity of the data in the SKU-110k dat
If you use the SKU-110k dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -113,7 +113,7 @@ The SKU-110k dataset consists of densely packed retail shelf images designed to
Training a YOLOv8 model on the SKU-110k dataset is straightforward. Here's an example to train a YOLOv8n model for 100 epochs with an image size of 640:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -165,7 +165,7 @@ These features make the SKU-110k dataset particularly valuable for training and
If you use the SKU-110k dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/detect/visdrone.md b/docs/en/datasets/detect/visdrone.md
index 4473bd2d93..c1060e9989 100644
--- a/docs/en/datasets/detect/visdrone.md
+++ b/docs/en/datasets/detect/visdrone.md
@@ -39,7 +39,7 @@ The VisDrone dataset is widely used for training and evaluating deep learning mo
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the Visdrone dataset, the `VisDrone.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/VisDrone.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/VisDrone.yaml).
-!!! Example "ultralytics/cfg/datasets/VisDrone.yaml"
+!!! example "ultralytics/cfg/datasets/VisDrone.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/VisDrone.yaml"
@@ -49,7 +49,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the VisDrone dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -84,7 +84,7 @@ The example showcases the variety and complexity of the data in the VisDrone dat
If you use the VisDrone dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -117,7 +117,7 @@ The [VisDrone Dataset](https://github.com/VisDrone/VisDrone-Dataset) is a large-
To train a YOLOv8 model on the VisDrone dataset for 100 epochs with an image size of 640, you can follow these steps:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -161,7 +161,7 @@ The configuration file for the VisDrone dataset, `VisDrone.yaml`, can be found i
If you use the VisDrone dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/detect/voc.md b/docs/en/datasets/detect/voc.md
index 9efb527990..0afc920529 100644
--- a/docs/en/datasets/detect/voc.md
+++ b/docs/en/datasets/detect/voc.md
@@ -31,7 +31,7 @@ The VOC dataset is widely used for training and evaluating deep learning models
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the VOC dataset, the `VOC.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/VOC.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/VOC.yaml).
-!!! Example "ultralytics/cfg/datasets/VOC.yaml"
+!!! example "ultralytics/cfg/datasets/VOC.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/VOC.yaml"
@@ -41,7 +41,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n model on the VOC dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -76,7 +76,7 @@ The example showcases the variety and complexity of the images in the VOC datase
If you use the VOC dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -103,7 +103,7 @@ The [PASCAL VOC](http://host.robots.ox.ac.uk/pascal/VOC/) (Visual Object Classes
To train a YOLOv8 model with the VOC dataset, you need the dataset configuration in a YAML file. Here's an example to start training a YOLOv8n model for 100 epochs with an image size of 640:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/detect/xview.md b/docs/en/datasets/detect/xview.md
index 53afa1957a..e7e2f3d3f7 100644
--- a/docs/en/datasets/detect/xview.md
+++ b/docs/en/datasets/detect/xview.md
@@ -34,7 +34,7 @@ The xView dataset is widely used for training and evaluating deep learning model
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the xView dataset, the `xView.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/xView.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/xView.yaml).
-!!! Example "ultralytics/cfg/datasets/xView.yaml"
+!!! example "ultralytics/cfg/datasets/xView.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/xView.yaml"
@@ -44,7 +44,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a model on the xView dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -79,7 +79,7 @@ The example showcases the variety and complexity of the data in the xView datase
If you use the xView dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -106,7 +106,7 @@ The [xView](http://xviewdataset.org/) dataset is one of the largest publicly ava
To train a model on the xView dataset using Ultralytics YOLO, follow these steps:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -147,7 +147,7 @@ The xView dataset comprises high-resolution satellite images collected from Worl
If you utilize the xView dataset in your research, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/explorer/api.md b/docs/en/datasets/explorer/api.md
index 87a09d19cb..905142efa7 100644
--- a/docs/en/datasets/explorer/api.md
+++ b/docs/en/datasets/explorer/api.md
@@ -48,7 +48,7 @@ dataframe = explorer.get_similar(img="path/to/image.jpg")
dataframe = explorer.get_similar(idx=0)
```
-!!! Tip "Note"
+!!! note
Embeddings table for a given dataset and model pair is only created once and reused. These use [LanceDB](https://lancedb.github.io/lancedb/) under the hood, which scales on-disk, so you can create and reuse embeddings for large datasets like COCO without running out of memory.
@@ -67,7 +67,7 @@ In case of multiple inputs, the aggregate of their embeddings is used.
You get a pandas dataframe with the `limit` number of most similar data points to the input, along with their distance in the embedding space. You can use this dataset to perform further filtering
-!!! Example "Semantic Search"
+!!! example "Semantic Search"
=== "Using Images"
@@ -110,7 +110,7 @@ You get a pandas dataframe with the `limit` number of most similar data points t
You can also plot the similar images using the `plot_similar` method. This method takes the same arguments as `get_similar` and plots the similar images in a grid.
-!!! Example "Plotting Similar Images"
+!!! example "Plotting Similar Images"
=== "Using Images"
@@ -143,7 +143,7 @@ You can also plot the similar images using the `plot_similar` method. This metho
This allows you to write how you want to filter your dataset using natural language. You don't have to be proficient in writing SQL queries. Our AI powered query generator will automatically do that under the hood. For example - you can say - "show me 100 images with exactly one person and 2 dogs. There can be other objects too" and it'll internally generate the query and show you those results.
Note: This works using LLMs under the hood so the results are probabilistic and might get things wrong sometimes
-!!! Example "Ask AI"
+!!! example "Ask AI"
```python
from ultralytics import Explorer
@@ -165,7 +165,7 @@ Note: This works using LLMs under the hood so the results are probabilistic and
You can run SQL queries on your dataset using the `sql_query` method. This method takes a SQL query as input and returns a pandas dataframe with the results.
-!!! Example "SQL Query"
+!!! example "SQL Query"
```python
from ultralytics import Explorer
@@ -182,7 +182,7 @@ You can run SQL queries on your dataset using the `sql_query` method. This metho
You can also plot the results of a SQL query using the `plot_sql_query` method. This method takes the same arguments as `sql_query` and plots the results in a grid.
-!!! Example "Plotting SQL Query Results"
+!!! example "Plotting SQL Query Results"
```python
from ultralytics import Explorer
@@ -199,7 +199,9 @@ You can also plot the results of a SQL query using the `plot_sql_query` method.
You can also work with the embeddings table directly. Once the embeddings table is created, you can access it using the `Explorer.table`
-!!! Tip "Explorer works on [LanceDB](https://lancedb.github.io/lancedb/) tables internally. You can access this table directly, using `Explorer.table` object and run raw queries, push down pre- and post-filters, etc."
+!!! tip
+
+ Explorer works on [LanceDB](https://lancedb.github.io/lancedb/) tables internally. You can access this table directly, using `Explorer.table` object and run raw queries, push down pre- and post-filters, etc.
```python
from ultralytics import Explorer
@@ -213,7 +215,7 @@ Here are some examples of what you can do with the table:
### Get raw Embeddings
-!!! Example
+!!! example
```python
from ultralytics import Explorer
@@ -228,7 +230,7 @@ Here are some examples of what you can do with the table:
### Advanced Querying with pre- and post-filters
-!!! Example
+!!! example
```python
from ultralytics import Explorer
@@ -270,11 +272,11 @@ It returns a pandas dataframe with the following columns:
- `count`: Number of images in the dataset that are closer than `max_dist` to the current image
- `sim_im_files`: List of paths to the `count` similar images
-!!! Tip
+!!! tip
For a given dataset, model, `max_dist` & `top_k` the similarity index once generated will be reused. In case, your dataset has changed, or you simply need to regenerate the similarity index, you can pass `force=True`.
-!!! Example "Similarity Index"
+!!! example "Similarity Index"
```python
from ultralytics import Explorer
diff --git a/docs/en/datasets/index.md b/docs/en/datasets/index.md
index 97603b42f8..f1e154ef7f 100644
--- a/docs/en/datasets/index.md
+++ b/docs/en/datasets/index.md
@@ -127,7 +127,7 @@ Contributing a new dataset involves several steps to ensure that it aligns well
### Example Code to Optimize and Zip a Dataset
-!!! Example "Optimize and Zip a Dataset"
+!!! example "Optimize and Zip a Dataset"
=== "Python"
@@ -205,7 +205,7 @@ Discover more about YOLO on the [Ultralytics YOLO](https://www.ultralytics.com/y
To optimize and zip a dataset using Ultralytics tools, follow this example code:
-!!! Example "Optimize and Zip a Dataset"
+!!! example "Optimize and Zip a Dataset"
=== "Python"
diff --git a/docs/en/datasets/obb/dota-v2.md b/docs/en/datasets/obb/dota-v2.md
index 7de8209a2b..70cac23cd2 100644
--- a/docs/en/datasets/obb/dota-v2.md
+++ b/docs/en/datasets/obb/dota-v2.md
@@ -60,7 +60,7 @@ DOTA serves as a benchmark for training and evaluating models specifically tailo
Typically, datasets incorporate a YAML (Yet Another Markup Language) file detailing the dataset's configuration. For DOTA v1 and DOTA v1.5, Ultralytics provides `DOTAv1.yaml` and `DOTAv1.5.yaml` files. For additional details on these as well as DOTA v2 please consult DOTA's official repository and documentation.
-!!! Example "DOTAv1.yaml"
+!!! example "DOTAv1.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/DOTAv1.yaml"
@@ -70,7 +70,7 @@ Typically, datasets incorporate a YAML (Yet Another Markup Language) file detail
To train DOTA dataset, we split original DOTA images with high-resolution into images with 1024x1024 resolution in multiscale way.
-!!! Example "Split images"
+!!! example "Split images"
=== "Python"
@@ -97,11 +97,11 @@ To train DOTA dataset, we split original DOTA images with high-resolution into i
To train a model on the DOTA v1 dataset, you can utilize the following code snippets. Always refer to your model's documentation for a thorough list of available arguments.
-!!! Warning
+!!! warning
Please note that all images and associated annotations in the DOTAv1 dataset can be used for academic purposes, but commercial use is prohibited. Your understanding and respect for the dataset creators' wishes are greatly appreciated!
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -136,7 +136,7 @@ The dataset's richness offers invaluable insights into object detection challeng
For those leveraging DOTA in their endeavors, it's pertinent to cite the relevant research papers:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -169,7 +169,7 @@ DOTA utilizes Oriented Bounding Boxes (OBB) for annotation, which are represente
To train a model on the DOTA dataset, you can use the following example with Ultralytics YOLO:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -204,7 +204,7 @@ For a detailed comparison and additional specifics, check the [dataset versions
DOTA images, which can be very large, are split into smaller resolutions for manageable training. Here's a Python snippet to split images:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/obb/dota8.md b/docs/en/datasets/obb/dota8.md
index 5a8bb29535..0188c31785 100644
--- a/docs/en/datasets/obb/dota8.md
+++ b/docs/en/datasets/obb/dota8.md
@@ -16,7 +16,7 @@ This dataset is intended for use with Ultralytics [HUB](https://hub.ultralytics.
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the DOTA8 dataset, the `dota8.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/dota8.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/dota8.yaml).
-!!! Example "ultralytics/cfg/datasets/dota8.yaml"
+!!! example "ultralytics/cfg/datasets/dota8.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/dota8.yaml"
@@ -26,7 +26,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n-obb model on the DOTA8 dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -61,7 +61,7 @@ The example showcases the variety and complexity of the images in the DOTA8 data
If you use the DOTA dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -90,7 +90,7 @@ The DOTA8 dataset is a small, versatile oriented object detection dataset made u
To train a YOLOv8n-obb model on the DOTA8 dataset for 100 epochs with an image size of 640, you can use the following code snippets. For comprehensive argument options, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/obb/index.md b/docs/en/datasets/obb/index.md
index d22d112172..02cd08e311 100644
--- a/docs/en/datasets/obb/index.md
+++ b/docs/en/datasets/obb/index.md
@@ -32,7 +32,7 @@ An example of a `*.txt` label file for the above image, which contains an object
To train a model using these OBB formats:
-!!! Example
+!!! example
=== "Python"
@@ -70,7 +70,7 @@ For those looking to introduce their own datasets with oriented bounding boxes,
Transitioning labels from the DOTA dataset format to the YOLO OBB format can be achieved with this script:
-!!! Example
+!!! example
=== "Python"
@@ -106,7 +106,7 @@ This script will reformat your DOTA annotations into a YOLO-compatible format.
Training a YOLOv8 model with OBBs involves ensuring your dataset is in the YOLO OBB format and then using the Ultralytics API to train the model. Here's an example in both Python and CLI:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/pose/coco.md b/docs/en/datasets/pose/coco.md
index 02bdb3e1a0..8addbd96e3 100644
--- a/docs/en/datasets/pose/coco.md
+++ b/docs/en/datasets/pose/coco.md
@@ -43,7 +43,7 @@ The COCO-Pose dataset is specifically used for training and evaluating deep lear
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO-Pose dataset, the `coco-pose.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco-pose.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco-pose.yaml).
-!!! Example "ultralytics/cfg/datasets/coco-pose.yaml"
+!!! example "ultralytics/cfg/datasets/coco-pose.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco-pose.yaml"
@@ -53,7 +53,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n-pose model on the COCO-Pose dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -88,7 +88,7 @@ The example showcases the variety and complexity of the images in the COCO-Pose
If you use the COCO-Pose dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -115,7 +115,7 @@ The [COCO-Pose](https://cocodataset.org/#keypoints-2017) dataset is a specialize
Training a YOLOv8 model on the COCO-Pose dataset can be accomplished using either Python or CLI commands. For example, to train a YOLOv8n-pose model for 100 epochs with an image size of 640, you can follow the steps below:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/pose/coco8-pose.md b/docs/en/datasets/pose/coco8-pose.md
index eeb0c3ddc3..c5847e1129 100644
--- a/docs/en/datasets/pose/coco8-pose.md
+++ b/docs/en/datasets/pose/coco8-pose.md
@@ -16,7 +16,7 @@ This dataset is intended for use with Ultralytics [HUB](https://hub.ultralytics.
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO8-Pose dataset, the `coco8-pose.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8-pose.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8-pose.yaml).
-!!! Example "ultralytics/cfg/datasets/coco8-pose.yaml"
+!!! example "ultralytics/cfg/datasets/coco8-pose.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco8-pose.yaml"
@@ -26,7 +26,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n-pose model on the COCO8-Pose dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -61,7 +61,7 @@ The example showcases the variety and complexity of the images in the COCO8-Pose
If you use the COCO dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -88,7 +88,7 @@ The COCO8-Pose dataset is a small, versatile pose detection dataset that include
To train a YOLOv8n-pose model on the COCO8-Pose dataset for 100 epochs with an image size of 640, follow these examples:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/pose/index.md b/docs/en/datasets/pose/index.md
index 700713d75e..4288497174 100644
--- a/docs/en/datasets/pose/index.md
+++ b/docs/en/datasets/pose/index.md
@@ -64,7 +64,7 @@ The `train` and `val` fields specify the paths to the directories containing the
## Usage
-!!! Example
+!!! example
=== "Python"
@@ -126,7 +126,7 @@ If you have your own dataset and would like to use it for training pose estimati
Ultralytics provides a convenient conversion tool to convert labels from the popular COCO dataset format to YOLO format:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/datasets/pose/tiger-pose.md b/docs/en/datasets/pose/tiger-pose.md
index 2462cf108c..13f62232d6 100644
--- a/docs/en/datasets/pose/tiger-pose.md
+++ b/docs/en/datasets/pose/tiger-pose.md
@@ -29,7 +29,7 @@ This dataset is intended for use with [Ultralytics HUB](https://hub.ultralytics.
A YAML (Yet Another Markup Language) file serves as the means to specify the configuration details of a dataset. It encompasses crucial data such as file paths, class definitions, and other pertinent information. Specifically, for the `tiger-pose.yaml` file, you can check [Ultralytics Tiger-Pose Dataset Configuration File](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/tiger-pose.yaml).
-!!! Example "ultralytics/cfg/datasets/tiger-pose.yaml"
+!!! example "ultralytics/cfg/datasets/tiger-pose.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/tiger-pose.yaml"
@@ -39,7 +39,7 @@ A YAML (Yet Another Markup Language) file serves as the means to specify the con
To train a YOLOv8n-pose model on the Tiger-Pose dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -72,7 +72,7 @@ The example showcases the variety and complexity of the images in the Tiger-Pose
## Inference Example
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
@@ -107,7 +107,7 @@ The Ultralytics Tiger-Pose dataset is designed for pose estimation tasks, consis
To train a YOLOv8n-pose model on the Tiger-Pose dataset for 100 epochs with an image size of 640, use the following code snippets. For more details, visit the [Training](../../modes/train.md) page:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -137,7 +137,7 @@ The `tiger-pose.yaml` file is used to specify the configuration details of the T
To perform inference using a YOLOv8 model trained on the Tiger-Pose dataset, you can use the following code snippets. For a detailed guide, visit the [Prediction](../../modes/predict.md) page:
-!!! Example "Inference Example"
+!!! example "Inference Example"
=== "Python"
diff --git a/docs/en/datasets/segment/carparts-seg.md b/docs/en/datasets/segment/carparts-seg.md
index 621fe9f2ba..6283cae8ce 100644
--- a/docs/en/datasets/segment/carparts-seg.md
+++ b/docs/en/datasets/segment/carparts-seg.md
@@ -37,7 +37,7 @@ Carparts Segmentation finds applications in automotive quality control, auto rep
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the Package Segmentation dataset, the `carparts-seg.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/carparts-seg.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/carparts-seg.yaml).
-!!! Example "ultralytics/cfg/datasets/carparts-seg.yaml"
+!!! example "ultralytics/cfg/datasets/carparts-seg.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/carparts-seg.yaml"
@@ -47,7 +47,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train Ultralytics YOLOv8n model on the Carparts Segmentation dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -81,7 +81,7 @@ The Carparts Segmentation dataset includes a diverse array of images and videos
If you integrate the Carparts Segmentation dataset into your research or development projects, please make reference to the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -112,7 +112,7 @@ The [Roboflow Carparts Segmentation Dataset](https://universe.roboflow.com/gianm
To train a YOLOv8 model on the Carparts Segmentation dataset, you can follow these steps:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/segment/coco.md b/docs/en/datasets/segment/coco.md
index ad372675fa..5c6b56402d 100644
--- a/docs/en/datasets/segment/coco.md
+++ b/docs/en/datasets/segment/coco.md
@@ -41,7 +41,7 @@ COCO-Seg is widely used for training and evaluating deep learning models in inst
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO-Seg dataset, the `coco.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml).
-!!! Example "ultralytics/cfg/datasets/coco.yaml"
+!!! example "ultralytics/cfg/datasets/coco.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco.yaml"
@@ -51,7 +51,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n-seg model on the COCO-Seg dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -86,7 +86,7 @@ The example showcases the variety and complexity of the images in the COCO-Seg d
If you use the COCO-Seg dataset in your research or development work, please cite the original COCO paper and acknowledge the extension to COCO-Seg:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -113,7 +113,7 @@ The [COCO-Seg](https://cocodataset.org/#home) dataset is an extension of the ori
To train a YOLOv8n-seg model on the COCO-Seg dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a detailed list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/segment/coco8-seg.md b/docs/en/datasets/segment/coco8-seg.md
index 15128ed24e..fb505a5615 100644
--- a/docs/en/datasets/segment/coco8-seg.md
+++ b/docs/en/datasets/segment/coco8-seg.md
@@ -16,7 +16,7 @@ This dataset is intended for use with Ultralytics [HUB](https://hub.ultralytics.
A YAML (Yet Another Markup Language) file is used to define the dataset configuration. It contains information about the dataset's paths, classes, and other relevant information. In the case of the COCO8-Seg dataset, the `coco8-seg.yaml` file is maintained at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8-seg.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco8-seg.yaml).
-!!! Example "ultralytics/cfg/datasets/coco8-seg.yaml"
+!!! example "ultralytics/cfg/datasets/coco8-seg.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/coco8-seg.yaml"
@@ -26,7 +26,7 @@ A YAML (Yet Another Markup Language) file is used to define the dataset configur
To train a YOLOv8n-seg model on the COCO8-Seg dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -61,7 +61,7 @@ The example showcases the variety and complexity of the images in the COCO8-Seg
If you use the COCO dataset in your research or development work, please cite the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -88,7 +88,7 @@ The **COCO8-Seg dataset** is a compact instance segmentation dataset by Ultralyt
To train a **YOLOv8n-seg** model on the COCO8-Seg dataset for 100 epochs with an image size of 640, you can use Python or CLI commands. Here's a quick example:
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
diff --git a/docs/en/datasets/segment/crack-seg.md b/docs/en/datasets/segment/crack-seg.md
index 32113dfc6d..ed66d7cf94 100644
--- a/docs/en/datasets/segment/crack-seg.md
+++ b/docs/en/datasets/segment/crack-seg.md
@@ -26,7 +26,7 @@ Crack segmentation finds practical applications in infrastructure maintenance, a
A YAML (Yet Another Markup Language) file is employed to outline the configuration of the dataset, encompassing details about paths, classes, and other pertinent information. Specifically, for the Crack Segmentation dataset, the `crack-seg.yaml` file is managed and accessible at [https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/crack-seg.yaml](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/crack-seg.yaml).
-!!! Example "ultralytics/cfg/datasets/crack-seg.yaml"
+!!! example "ultralytics/cfg/datasets/crack-seg.yaml"
```yaml
--8<-- "ultralytics/cfg/datasets/crack-seg.yaml"
@@ -36,7 +36,7 @@ A YAML (Yet Another Markup Language) file is employed to outline the configurati
To train Ultralytics YOLOv8n model on the Crack Segmentation dataset for 100 epochs with an image size of 640, you can use the following code snippets. For a comprehensive list of available arguments, refer to the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -71,7 +71,7 @@ The Crack Segmentation dataset comprises a varied collection of images and video
If you incorporate the crack segmentation dataset into your research or development endeavors, kindly reference the following paper:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
@@ -102,7 +102,7 @@ The [Roboflow Crack Segmentation Dataset](https://universe.roboflow.com/universi
To train an Ultralytics YOLOv8 model on the Crack Segmentation dataset, use the following code snippets. Detailed instructions and further parameters can be found on the model [Training](../../modes/train.md) page.
-!!! Example "Train Example"
+!!! example "Train Example"
=== "Python"
@@ -135,7 +135,7 @@ Ultralytics YOLO offers advanced real-time object detection, segmentation, and c
If you incorporate the Crack Segmentation Dataset into your research, please use the following BibTeX reference:
-!!! Quote ""
+!!! quote ""
=== "BibTeX"
diff --git a/docs/en/datasets/segment/index.md b/docs/en/datasets/segment/index.md
index 09c9279715..838dda2c7a 100644
--- a/docs/en/datasets/segment/index.md
+++ b/docs/en/datasets/segment/index.md
@@ -33,7 +33,7 @@ Here is an example of the YOLO dataset format for a single image with two object
1 0.504 0.000 0.501 0.004 0.498 0.004 0.493 0.010 0.492 0.0104
```
-!!! Tip "Tip"
+!!! tip "Tip"
- The length of each row does **not** have to be equal.
- Each segmentation label must have a **minimum of 3 xy points**: `
-!!! Note
+!!! note
This guide has been tested with both [Seeed Studio reComputer J4012](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) which is based on NVIDIA Jetson Orin NX 16GB running JetPack release of [JP5.1.3](https://developer.nvidia.com/embedded/jetpack-sdk-513) and [Seeed Studio reComputer J1020 v2](https://www.seeedstudio.com/reComputer-J1020-v2-p-5498.html) which is based on NVIDIA Jetson Nano 4GB running JetPack release of [JP4.6.4](https://developer.nvidia.com/jetpack-sdk-464). It is expected to work across all the NVIDIA Jetson hardware lineup including latest and legacy.
@@ -39,7 +39,7 @@ Before you start to follow this guide:
- For JetPack 4.6.4, install [DeepStream 6.0.1](https://docs.nvidia.com/metropolis/deepstream/6.0.1/dev-guide/text/DS_Quickstart.html)
- For JetPack 5.1.3, install [DeepStream 6.3](https://docs.nvidia.com/metropolis/deepstream/6.3/dev-guide/text/DS_Quickstart.html)
-!!! Tip
+!!! tip
In this guide we have used the Debian package method of installing DeepStream SDK to the Jetson device. You can also visit the [DeepStream SDK on Jetson (Archived)](https://developer.nvidia.com/embedded/deepstream-on-jetson-downloads-archived) to access legacy versions of DeepStream.
@@ -67,7 +67,7 @@ Here we are using [marcoslucianops/DeepStream-Yolo](https://github.com/marcosluc
wget https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8s.pt
```
- !!! Note
+ !!! note
You can also use a [custom trained YOLOv8 model](https://docs.ultralytics.com/modes/train/).
@@ -77,7 +77,7 @@ Here we are using [marcoslucianops/DeepStream-Yolo](https://github.com/marcosluc
python3 utils/export_yoloV8.py -w yolov8s.pt
```
- !!! Note "Pass the below arguments to the above command"
+ !!! note "Pass the below arguments to the above command"
For DeepStream 6.0.1, use opset 12 or lower. The default opset is 16.
@@ -175,13 +175,13 @@ Here we are using [marcoslucianops/DeepStream-Yolo](https://github.com/marcosluc
deepstream-app -c deepstream_app_config.txt
```
-!!! Note
+!!! note
It will take a long time to generate the TensorRT engine file before starting the inference. So please be patient.
-!!! Note
+!!! note
This guide has been tested with both [Seeed Studio reComputer J4012](https://www.seeedstudio.com/reComputer-J4012-p-5586.html) which is based on NVIDIA Jetson Orin NX 16GB running the latest stable JetPack release of [JP6.0](https://developer.nvidia.com/embedded/jetpack-sdk-60), JetPack release of [JP5.1.3](https://developer.nvidia.com/embedded/jetpack-sdk-513) and [Seeed Studio reComputer J1020 v2](https://www.seeedstudio.com/reComputer-J1020-v2-p-5498.html) which is based on NVIDIA Jetson Nano 4GB running JetPack release of [JP4.6.1](https://developer.nvidia.com/embedded/jetpack-sdk-461). It is expected to work across all the NVIDIA Jetson hardware lineup including latest and legacy.
@@ -57,7 +57,7 @@ The first step after getting your hands on an NVIDIA Jetson device is to flash N
3. If you own a Seeed Studio reComputer J4012 device, you can [flash JetPack to the included SSD](https://wiki.seeedstudio.com/reComputer_J4012_Flash_Jetpack/) and if you own a Seeed Studio reComputer J1020 v2 device, you can [flash JetPack to the eMMC/ SSD](https://wiki.seeedstudio.com/reComputer_J2021_J202_Flash_Jetpack/).
4. If you own any other third party device powered by the NVIDIA Jetson module, it is recommended to follow [command-line flashing](https://docs.nvidia.com/jetson/archives/r35.5.0/DeveloperGuide/IN/QuickStart.html).
-!!! Note
+!!! note
For methods 3 and 4 above, after flashing the system and booting the device, please enter "sudo apt update && sudo apt install nvidia-jetpack -y" on the device terminal to install all the remaining JetPack components needed.
@@ -157,7 +157,7 @@ wget https://nvidia.box.com/shared/static/48dtuob7meiw6ebgfsfqakc9vse62sg4.whl -
pip install onnxruntime_gpu-1.18.0-cp310-cp310-linux_aarch64.whl
```
-!!! Note
+!!! note
`onnxruntime-gpu` will automatically revert back the numpy version to latest. So we need to reinstall numpy to `1.23.5` to fix an issue by executing:
@@ -230,7 +230,7 @@ wget https://nvidia.box.com/shared/static/zostg6agm00fb6t5uisw51qi6kpcuwzd.whl -
pip install onnxruntime_gpu-1.17.0-cp38-cp38-linux_aarch64.whl
```
-!!! Note
+!!! note
`onnxruntime-gpu` will automatically revert back the numpy version to latest. So we need to reinstall numpy to `1.23.5` to fix an issue by executing:
@@ -244,7 +244,7 @@ Out of all the model export formats supported by Ultralytics, TensorRT delivers
The YOLOv8n model in PyTorch format is converted to TensorRT to run inference with the exported model.
-!!! Example
+!!! example
=== "Python"
@@ -274,7 +274,7 @@ The YOLOv8n model in PyTorch format is converted to TensorRT to run inference wi
yolo predict model=yolov8n.engine source='https://ultralytics.com/images/bus.jpg'
```
-!!! Note
+!!! note
Visit the [Export page](../modes/export.md#arguments) to access additional arguments when exporting models to different model formats
@@ -294,7 +294,7 @@ Even though all model exports are working with NVIDIA Jetson, we have only inclu
The below table represents the benchmark results for five different models (YOLOv8n, YOLOv8s, YOLOv8m, YOLOv8l, YOLOv8x) across ten different formats (PyTorch, TorchScript, ONNX, OpenVINO, TensorRT, TF SavedModel, TF GraphDef, TF Lite, PaddlePaddle, NCNN), giving us the status, size, mAP50-95(B) metric, and inference time for each combination.
-!!! Performance
+!!! performance
=== "YOLOv8n"
@@ -377,7 +377,7 @@ The below table represents the benchmark results for five different models (YOLO
To reproduce the above Ultralytics benchmarks on all export [formats](../modes/export.md) run this code:
-!!! Example
+!!! example
=== "Python"
diff --git a/docs/en/guides/object-blurring.md b/docs/en/guides/object-blurring.md
index e6a21338da..48a4a04f2a 100644
--- a/docs/en/guides/object-blurring.md
+++ b/docs/en/guides/object-blurring.md
@@ -27,7 +27,7 @@ Object blurring with [Ultralytics YOLOv8](https://github.com/ultralytics/ultraly
- **Selective Focus**: YOLOv8 allows for selective blurring, enabling users to target specific objects, ensuring a balance between privacy and retaining relevant visual information.
- **Real-time Processing**: YOLOv8's efficiency enables object blurring in real-time, making it suitable for applications requiring on-the-fly privacy enhancements in dynamic environments.
-!!! Example "Object Blurring using YOLOv8 Example"
+!!! example "Object Blurring using YOLOv8 Example"
=== "Object Blurring"
diff --git a/docs/en/guides/object-counting.md b/docs/en/guides/object-counting.md
index 00aa917454..1204dfce51 100644
--- a/docs/en/guides/object-counting.md
+++ b/docs/en/guides/object-counting.md
@@ -46,7 +46,7 @@ Object counting with [Ultralytics YOLOv8](https://github.com/ultralytics/ultraly
|  |  |
| Conveyor Belt Packets Counting Using Ultralytics YOLOv8 | Fish Counting in Sea using Ultralytics YOLOv8 |
-!!! Example "Object Counting using YOLOv8 Example"
+!!! example "Object Counting using YOLOv8 Example"
=== "Count in Region"
diff --git a/docs/en/guides/object-cropping.md b/docs/en/guides/object-cropping.md
index 3efaba93e1..caf831fe88 100644
--- a/docs/en/guides/object-cropping.md
+++ b/docs/en/guides/object-cropping.md
@@ -34,7 +34,7 @@ Object cropping with [Ultralytics YOLOv8](https://github.com/ultralytics/ultraly
|  |
| Suitcases Cropping at airport conveyor belt using Ultralytics YOLOv8 |
-!!! Example "Object Cropping using YOLOv8 Example"
+!!! example "Object Cropping using YOLOv8 Example"
=== "Object Cropping"
diff --git a/docs/en/guides/parking-management.md b/docs/en/guides/parking-management.md
index e25936fbd3..bd5b0edd41 100644
--- a/docs/en/guides/parking-management.md
+++ b/docs/en/guides/parking-management.md
@@ -38,18 +38,18 @@ Parking management with [Ultralytics YOLOv8](https://github.com/ultralytics/ultr
### Selection of Points
-!!! Tip "Point Selection is now Easy"
+!!! tip "Point Selection is now Easy"
Choosing parking points is a critical and complex task in parking management systems. Ultralytics streamlines this process by providing a tool that lets you define parking lot areas, which can be utilized later for additional processing.
- Capture a frame from the video or camera stream where you want to manage the parking lot.
- Use the provided code to launch a graphical interface, where you can select an image and start outlining parking regions by mouse click to create polygons.
-!!! Warning "Image Size"
+!!! warning "Image Size"
Max Image Size of 1920 * 1080 supported
-!!! Example "Parking slots Annotator Ultralytics YOLOv8"
+!!! example "Parking slots Annotator Ultralytics YOLOv8"
=== "Parking Annotator"
@@ -65,7 +65,7 @@ Parking management with [Ultralytics YOLOv8](https://github.com/ultralytics/ultr
### Python Code for Parking Management
-!!! Example "Parking management using YOLOv8 Example"
+!!! example "Parking management using YOLOv8 Example"
=== "Parking Management"
diff --git a/docs/en/guides/queue-management.md b/docs/en/guides/queue-management.md
index ad599770b2..7abc31d31c 100644
--- a/docs/en/guides/queue-management.md
+++ b/docs/en/guides/queue-management.md
@@ -33,7 +33,7 @@ Queue management using [Ultralytics YOLOv8](https://github.com/ultralytics/ultra
|  |  |
| Queue management at airport ticket counter Using Ultralytics YOLOv8 | Queue monitoring in crowd Ultralytics YOLOv8 |
-!!! Example "Queue Management using YOLOv8 Example"
+!!! example "Queue Management using YOLOv8 Example"
=== "Queue Manager"
diff --git a/docs/en/guides/raspberry-pi.md b/docs/en/guides/raspberry-pi.md
index 997c08547b..674bc847b3 100644
--- a/docs/en/guides/raspberry-pi.md
+++ b/docs/en/guides/raspberry-pi.md
@@ -19,7 +19,7 @@ This comprehensive guide provides a detailed walkthrough for deploying Ultralyti
Watch: Raspberry Pi 5 updates and improvements.

@@ -56,7 +56,7 @@ Ultralytics provides various installation methods including pip, conda, and Dock conda install -c conda-forge ultralytics ``` - !!! Note + !!! note If you are installing in a CUDA environment best practice is to install `ultralytics`, `pytorch` and `pytorch-cuda` in the same command to allow the conda package manager to resolve any conflicts, or else to install `pytorch-cuda` last to allow it override the CPU-specific `pytorch` package if necessary. ```bash @@ -141,7 +141,7 @@ Ultralytics provides various installation methods including pip, conda, and Dock See the `ultralytics` [pyproject.toml](https://github.com/ultralytics/ultralytics/blob/main/pyproject.toml) file for a list of dependencies. Note that all examples above install all required dependencies. -!!! Tip "Tip" +!!! tip "Tip" PyTorch requirements vary by operating system and CUDA requirements, so it's recommended to install PyTorch first following instructions at [https://pytorch.org/get-started/locally](https://pytorch.org/get-started/locally/). @@ -153,7 +153,7 @@ See the `ultralytics` [pyproject.toml](https://github.com/ultralytics/ultralytic The Ultralytics command line interface (CLI) allows for simple single-line commands without the need for a Python environment. CLI requires no customization or Python code. You can simply run all tasks from the terminal with the `yolo` command. Check out the [CLI Guide](usage/cli.md) to learn more about using YOLOv8 from the command line. -!!! Example +!!! example === "Syntax" @@ -208,7 +208,7 @@ The Ultralytics command line interface (CLI) allows for simple single-line comma yolo cfg ``` -!!! Warning "Warning" +!!! warning "Warning" Arguments must be passed as `arg=val` pairs, split by an equals `=` sign and delimited by spaces between pairs. Do not use `--` argument prefixes or commas `,` between arguments. @@ -225,7 +225,7 @@ YOLOv8's Python interface allows for seamless integration into your Python proje For example, users can load a model, train it, evaluate its performance on a validation set, and even export it to ONNX format with just a few lines of code. Check out the [Python Guide](usage/python.md) to learn more about using YOLOv8 within your Python projects. -!!! Example +!!! example ```python from ultralytics import YOLO @@ -259,7 +259,7 @@ The Ultralytics library provides a powerful settings management system to enable To gain insight into the current configuration of your settings, you can view them directly: -!!! Example "View settings" +!!! example "View settings" === "Python" @@ -285,7 +285,7 @@ To gain insight into the current configuration of your settings, you can view th Ultralytics allows users to easily modify their settings. Changes can be performed in the following ways: -!!! Example "Update settings" +!!! example "Update settings" === "Python" diff --git a/docs/en/tasks/classify.md b/docs/en/tasks/classify.md index 6fa0fabbfe..07a9cc4ecb 100644 --- a/docs/en/tasks/classify.md +++ b/docs/en/tasks/classify.md @@ -24,7 +24,7 @@ The output of an image classifier is a single class label and a confidence score Watch: Explore Ultralytics YOLO Tasks: Image Classification using Ultralytics HUB
-!!! Tip "Tip" +!!! tip YOLOv8 Classify models use the `-cls` suffix, i.e. `yolov8n-cls.pt` and are pretrained on [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml). @@ -49,7 +49,7 @@ YOLOv8 pretrained Classify models are shown here. Detect, Segment and Pose model Train YOLOv8n-cls on the MNIST160 dataset for 100 epochs at image size 64. For a full list of available arguments see the [Configuration](../usage/cfg.md) page. -!!! Example +!!! example === "Python" @@ -86,7 +86,7 @@ YOLO classification dataset format can be found in detail in the [Dataset Guide] Validate trained YOLOv8n-cls model accuracy on the MNIST160 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example +!!! example === "Python" @@ -114,7 +114,7 @@ Validate trained YOLOv8n-cls model accuracy on the MNIST160 dataset. No argument Use a trained YOLOv8n-cls model to run predictions on images. -!!! Example +!!! example === "Python" @@ -142,7 +142,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page. Export a YOLOv8n-cls model to a different format like ONNX, CoreML, etc. -!!! Example +!!! example === "Python" @@ -180,7 +180,7 @@ YOLOv8 models, such as `yolov8n-cls.pt`, are designed for efficient image classi To train a YOLOv8 model, you can use either Python or CLI commands. For example, to train a `yolov8n-cls` model on the MNIST160 dataset for 100 epochs at an image size of 64: -!!! Example +!!! example === "Python" @@ -210,7 +210,7 @@ Pretrained YOLOv8 classification models can be found in the [Models](https://git You can export a trained YOLOv8 model to various formats using Python or CLI commands. For instance, to export a model to ONNX format: -!!! Example +!!! example === "Python" @@ -236,7 +236,7 @@ For detailed export options, refer to the [Export](../modes/export.md) page. To validate a trained model's accuracy on a dataset like MNIST160, you can use the following Python or CLI commands: -!!! Example +!!! example === "Python" diff --git a/docs/en/tasks/detect.md b/docs/en/tasks/detect.md index d5b4e1f0dc..d3542f184d 100644 --- a/docs/en/tasks/detect.md +++ b/docs/en/tasks/detect.md @@ -23,7 +23,7 @@ The output of an object detector is a set of bounding boxes that enclose the obj Watch: Object Detection with Pre-trained Ultralytics YOLOv8 Model. -!!! Tip "Tip" +!!! tip YOLOv8 Detect models are the default YOLOv8 models, i.e. `yolov8n.pt` and are pretrained on [COCO](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml). @@ -48,7 +48,7 @@ YOLOv8 pretrained Detect models are shown here. Detect, Segment and Pose models Train YOLOv8n on the COCO8 dataset for 100 epochs at image size 640. For a full list of available arguments see the [Configuration](../usage/cfg.md) page. -!!! Example +!!! example === "Python" @@ -85,7 +85,7 @@ YOLO detection dataset format can be found in detail in the [Dataset Guide](../d Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example +!!! example === "Python" @@ -115,7 +115,7 @@ Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need t Use a trained YOLOv8n model to run predictions on images. -!!! Example +!!! example === "Python" @@ -143,7 +143,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page. Export a YOLOv8n model to a different format like ONNX, CoreML, etc. -!!! Example +!!! example === "Python" @@ -181,7 +181,7 @@ Training a YOLOv8 model on a custom dataset involves a few steps: 2. **Load the Model**: Use the Ultralytics YOLO library to load a pre-trained model or create a new model from a YAML file. 3. **Train the Model**: Execute the `train` method in Python or the `yolo detect train` command in CLI. -!!! Example +!!! example === "Python" @@ -219,7 +219,7 @@ For a detailed list and performance metrics, refer to the [Models](https://githu To validate the accuracy of your trained YOLOv8 model, you can use the `.val()` method in Python or the `yolo detect val` command in CLI. This will provide metrics like mAP50-95, mAP50, and more. -!!! Example +!!! example === "Python" @@ -246,7 +246,7 @@ For more validation details, visit the [Val](../modes/val.md) page. Ultralytics YOLOv8 allows exporting models to various formats such as ONNX, TensorRT, CoreML, and more to ensure compatibility across different platforms and devices. -!!! Example +!!! example === "Python" diff --git a/docs/en/tasks/index.md b/docs/en/tasks/index.md index e52fd53700..5b7b75afd6 100644 --- a/docs/en/tasks/index.md +++ b/docs/en/tasks/index.md @@ -76,7 +76,7 @@ To use Ultralytics YOLOv8 for object detection, follow these steps: 2. Train the YOLOv8 model using the detection task. 3. Use the model to make predictions by feeding in new images or video frames. -!!! Example +!!! example === "Python" diff --git a/docs/en/tasks/obb.md b/docs/en/tasks/obb.md index d49a289beb..a603ab1a84 100644 --- a/docs/en/tasks/obb.md +++ b/docs/en/tasks/obb.md @@ -15,7 +15,7 @@ The output of an oriented object detector is a set of rotated bounding boxes tha -!!! Tip "Tip" +!!! tip YOLOv8 OBB models use the `-obb` suffix, i.e. `yolov8n-obb.pt` and are pretrained on [DOTAv1](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/DOTAv1.yaml). @@ -69,7 +69,7 @@ YOLOv8 pretrained OBB models are shown here, which are pretrained on the [DOTAv1 Train YOLOv8n-obb on the `dota8.yaml` dataset for 100 epochs at image size 640. For a full list of available arguments see the [Configuration](../usage/cfg.md) page. -!!! Example +!!! example === "Python" @@ -107,7 +107,7 @@ OBB dataset format can be found in detail in the [Dataset Guide](../datasets/obb Validate trained YOLOv8n-obb model accuracy on the DOTA8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example +!!! example === "Python" @@ -137,7 +137,7 @@ retains its training `data` and arguments as model attributes. Use a trained YOLOv8n-obb model to run predictions on images. -!!! Example +!!! example === "Python" @@ -165,7 +165,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page. Export a YOLOv8n-obb model to a different format like ONNX, CoreML, etc. -!!! Example +!!! example === "Python" @@ -203,7 +203,7 @@ Oriented Bounding Boxes (OBB) include an additional angle to enhance object loca To train a YOLOv8n-obb model with a custom dataset, follow the example below using Python or CLI: -!!! Example +!!! example === "Python" @@ -233,7 +233,7 @@ YOLOv8-OBB models are pretrained on datasets like [DOTAv1](https://github.com/ul Exporting a YOLOv8-OBB model to ONNX format is straightforward using either Python or CLI: -!!! Example +!!! example === "Python" @@ -259,7 +259,7 @@ For more export formats and details, refer to the [Export](../modes/export.md) p To validate a YOLOv8n-obb model, you can use Python or CLI commands as shown below: -!!! Example +!!! example === "Python" diff --git a/docs/en/tasks/pose.md b/docs/en/tasks/pose.md index ffa0a39ffb..81384cc951 100644 --- a/docs/en/tasks/pose.md +++ b/docs/en/tasks/pose.md @@ -36,7 +36,7 @@ The output of a pose estimation model is a set of points that represent the keyp -!!! Tip "Tip" +!!! tip YOLOv8 _pose_ models use the `-pose` suffix, i.e. `yolov8n-pose.pt`. These models are trained on the [COCO keypoints](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco-pose.yaml) dataset and are suitable for a variety of pose estimation tasks. @@ -82,7 +82,7 @@ YOLOv8 pretrained Pose models are shown here. Detect, Segment and Pose models ar Train a YOLOv8-pose model on the COCO128-pose dataset. -!!! Example +!!! example === "Python" @@ -120,7 +120,7 @@ YOLO pose dataset format can be found in detail in the [Dataset Guide](../datase Validate trained YOLOv8n-pose model accuracy on the COCO128-pose dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example +!!! example === "Python" @@ -150,7 +150,7 @@ retains its training `data` and arguments as model attributes. Use a trained YOLOv8n-pose model to run predictions on images. -!!! Example +!!! example === "Python" @@ -178,7 +178,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page. Export a YOLOv8n Pose model to a different format like ONNX, CoreML, etc. -!!! Example +!!! example === "Python" diff --git a/docs/en/tasks/segment.md b/docs/en/tasks/segment.md index 96090fb2ee..56a3b3ba24 100644 --- a/docs/en/tasks/segment.md +++ b/docs/en/tasks/segment.md @@ -24,7 +24,7 @@ The output of an instance segmentation model is a set of masks or contours that Watch: Run Segmentation with Pre-Trained Ultralytics YOLOv8 Model in Python. -!!! Tip "Tip" +!!! tip YOLOv8 Segment models use the `-seg` suffix, i.e. `yolov8n-seg.pt` and are pretrained on [COCO](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml). @@ -49,7 +49,7 @@ YOLOv8 pretrained Segment models are shown here. Detect, Segment and Pose models Train YOLOv8n-seg on the COCO128-seg dataset for 100 epochs at image size 640. For a full list of available arguments see the [Configuration](../usage/cfg.md) page. -!!! Example +!!! example === "Python" @@ -87,7 +87,7 @@ YOLO segmentation dataset format can be found in detail in the [Dataset Guide](. Validate trained YOLOv8n-seg model accuracy on the COCO128-seg dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example +!!! example === "Python" @@ -121,7 +121,7 @@ retains its training `data` and arguments as model attributes. Use a trained YOLOv8n-seg model to run predictions on images. -!!! Example +!!! example === "Python" @@ -149,7 +149,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page. Export a YOLOv8n-seg model to a different format like ONNX, CoreML, etc. -!!! Example +!!! example === "Python" @@ -183,7 +183,7 @@ See full `export` details in the [Export](../modes/export.md) page. To train a YOLOv8 segmentation model on a custom dataset, you first need to prepare your dataset in the YOLO segmentation format. You can use tools like [JSON2YOLO](https://github.com/ultralytics/JSON2YOLO) to convert datasets from other formats. Once your dataset is ready, you can train the model using Python or CLI commands: -!!! Example +!!! example === "Python" @@ -217,7 +217,7 @@ Ultralytics YOLOv8 is a state-of-the-art model recognized for its high accuracy Loading and validating a pretrained YOLOv8 segmentation model is straightforward. Here's how you can do it using both Python and CLI: -!!! Example +!!! example === "Python" @@ -245,7 +245,7 @@ These steps will provide you with validation metrics like Mean Average Precision Exporting a YOLOv8 segmentation model to ONNX format is simple and can be done using Python or CLI commands: -!!! Example +!!! example === "Python" diff --git a/docs/en/usage/cfg.md b/docs/en/usage/cfg.md index 299c8acd7a..0f8776cd21 100644 --- a/docs/en/usage/cfg.md +++ b/docs/en/usage/cfg.md @@ -19,7 +19,7 @@ YOLO settings and hyperparameters play a critical role in the model's performanc Ultralytics commands use the following syntax: -!!! Example +!!! example === "CLI" diff --git a/docs/en/usage/cli.md b/docs/en/usage/cli.md index ba649cbf82..d2ba49277c 100644 --- a/docs/en/usage/cli.md +++ b/docs/en/usage/cli.md @@ -19,7 +19,7 @@ The YOLO command line interface (CLI) allows for simple single-line commands wit Watch: Mastering Ultralytics YOLOv8: CLI -!!! Example +!!! example === "Syntax" @@ -79,7 +79,7 @@ Where: - `MODE` (required) is one of `[train, val, predict, export, track, benchmark]` - `ARGS` (optional) are any number of custom `arg=value` pairs like `imgsz=320` that override defaults. For a full list of available `ARGS` see the [Configuration](cfg.md) page and `defaults.yaml` -!!! Warning "Warning" +!!! warning "Warning" Arguments must be passed as `arg=val` pairs, split by an equals `=` sign and delimited by spaces ` ` between pairs. Do not use `--` argument prefixes or commas `,` between arguments. @@ -91,7 +91,7 @@ Where: Train YOLOv8n on the COCO8 dataset for 100 epochs at image size 640. For a full list of available arguments see the [Configuration](cfg.md) page. -!!! Example "Example" +!!! example "Example" === "Train" @@ -111,7 +111,7 @@ Train YOLOv8n on the COCO8 dataset for 100 epochs at image size 640. For a full Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. -!!! Example "Example" +!!! example "Example" === "Official" @@ -131,7 +131,7 @@ Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need t Use a trained YOLOv8n model to run predictions on images. -!!! Example "Example" +!!! example "Example" === "Official" @@ -151,7 +151,7 @@ Use a trained YOLOv8n model to run predictions on images. Export a YOLOv8n model to a different format like ONNX, CoreML, etc. -!!! Example "Example" +!!! example "Example" === "Official" @@ -177,7 +177,7 @@ See full `export` details in the [Export](../modes/export.md) page. Default arguments can be overridden by simply passing them as arguments in the CLI in `arg=value` pairs. -!!! Tip "" +!!! tip "" === "Train" @@ -208,7 +208,7 @@ To do this first create a copy of `default.yaml` in your current working dir wit This will create `default_copy.yaml`, which you can then pass as `cfg=default_copy.yaml` along with any additional args, like `imgsz=320` in this example: -!!! Example +!!! example === "CLI" diff --git a/docs/en/usage/python.md b/docs/en/usage/python.md index 90b2fcfdac..66b2e6c61d 100644 --- a/docs/en/usage/python.md +++ b/docs/en/usage/python.md @@ -21,7 +21,7 @@ Welcome to the YOLOv8 Python Usage documentation! This guide is designed to help For example, users can load a model, train it, evaluate its performance on a validation set, and even export it to ONNX format with just a few lines of code. -!!! Example "Python" +!!! example "Python" ```python from ultralytics import YOLO @@ -49,7 +49,7 @@ For example, users can load a model, train it, evaluate its performance on a val Train mode is used for training a YOLOv8 model on a custom dataset. In this mode, the model is trained using the specified dataset and hyperparameters. The training process involves optimizing the model's parameters so that it can accurately predict the classes and locations of objects in an image. -!!! Example "Train" +!!! example "Train" === "From pretrained(recommended)" @@ -82,7 +82,7 @@ Train mode is used for training a YOLOv8 model on a custom dataset. In this mode Val mode is used for validating a YOLOv8 model after it has been trained. In this mode, the model is evaluated on a validation set to measure its accuracy and generalization performance. This mode can be used to tune the hyperparameters of the model to improve its performance. -!!! Example "Val" +!!! example "Val" === "Val after training" @@ -120,7 +120,7 @@ Val mode is used for validating a YOLOv8 model after it has been trained. In thi Predict mode is used for making predictions using a trained YOLOv8 model on new images or videos. In this mode, the model is loaded from a checkpoint file, and the user can provide images or videos to perform inference. The model predicts the classes and locations of objects in the input images or videos. -!!! Example "Predict" +!!! example "Predict" === "From source" @@ -191,7 +191,7 @@ Predict mode is used for making predictions using a trained YOLOv8 model on new Export mode is used for exporting a YOLOv8 model to a format that can be used for deployment. In this mode, the model is converted to a format that can be used by other software applications or hardware devices. This mode is useful when deploying the model to production environments. -!!! Example "Export" +!!! example "Export" === "Export to ONNX" @@ -219,7 +219,7 @@ Export mode is used for exporting a YOLOv8 model to a format that can be used fo Track mode is used for tracking objects in real-time using a YOLOv8 model. In this mode, the model is loaded from a checkpoint file, and the user can provide a live video stream to perform real-time object tracking. This mode is useful for applications such as surveillance systems or self-driving cars. -!!! Example "Track" +!!! example "Track" === "Python" @@ -242,7 +242,7 @@ Track mode is used for tracking objects in real-time using a YOLOv8 model. In th Benchmark mode is used to profile the speed and accuracy of various export formats for YOLOv8. The benchmarks provide information on the size of the exported format, its `mAP50-95` metrics (for object detection and segmentation) or `accuracy_top5` metrics (for classification), and the inference time in milliseconds per image across various export formats like ONNX, OpenVINO, TensorRT and others. This information can help users choose the optimal export format for their specific use case based on their requirements for speed and accuracy. -!!! Example "Benchmark" +!!! example "Benchmark" === "Python" @@ -260,7 +260,7 @@ Benchmark mode is used to profile the speed and accuracy of various export forma Explorer API can be used to explore datasets with advanced semantic, vector-similarity and SQL search among other features. It also enabled searching for images based on their content using natural language by utilizing the power of LLMs. The Explorer API allows you to write your own dataset exploration notebooks or scripts to get insights into your datasets. -!!! Example "Semantic Search Using Explorer" +!!! example "Semantic Search Using Explorer" === "Using Images" @@ -304,7 +304,7 @@ Explorer API can be used to explore datasets with advanced semantic, vector-simi `YOLO` model class is a high-level wrapper on the Trainer classes. Each YOLO task has its own trainer that inherits from `BaseTrainer`. -!!! Tip "Detection Trainer Example" +!!! tip "Detection Trainer Example" ```python from ultralytics.models.yolo import DetectionPredictor, DetectionTrainer, DetectionValidator diff --git a/docs/en/yolov5/tutorials/roboflow_datasets_integration.md b/docs/en/yolov5/tutorials/roboflow_datasets_integration.md index ac1454b201..3047e48c2e 100644 --- a/docs/en/yolov5/tutorials/roboflow_datasets_integration.md +++ b/docs/en/yolov5/tutorials/roboflow_datasets_integration.md @@ -8,7 +8,7 @@ keywords: Roboflow, YOLOv5, data management, dataset labeling, dataset versionin You can now use Roboflow to organize, label, prepare, version, and host your datasets for training YOLOv5 🚀 models. Roboflow is free to use with YOLOv5 if you make your workspace public. -!!! Question "Licensing" +!!! question "Licensing" Ultralytics offers two licensing options: diff --git a/docs/en/yolov5/tutorials/train_custom_data.md b/docs/en/yolov5/tutorials/train_custom_data.md index 69888785cd..f7cc6f098c 100644 --- a/docs/en/yolov5/tutorials/train_custom_data.md +++ b/docs/en/yolov5/tutorials/train_custom_data.md @@ -25,7 +25,7 @@ pip install -r requirements.txt # install Creating a custom model to detect your objects is an iterative process of collecting and organizing images, labeling your objects of interest, training a model, deploying it into the wild to make predictions, and then using that deployed model to collect examples of edge cases to repeat and improve. -!!! Question "Licensing" +!!! question "Licensing" Ultralytics offers two licensing options: @@ -137,11 +137,11 @@ Train a YOLOv5s model on COCO128 by specifying dataset, batch-size, image size a python train.py --img 640 --epochs 3 --data coco128.yaml --weights yolov5s.pt ``` -!!! Tip "Tip" +!!! tip "Tip" 💡 Add `--cache ram` or `--cache disk` to speed up training (requires significant RAM/disk resources). -!!! Tip "Tip" +!!! tip "Tip" 💡 Always train from a local dataset. Mounted or network drives like Google Drive will be very slow. diff --git a/docs/mkdocs_github_authors.yaml b/docs/mkdocs_github_authors.yaml index 1f47cd8200..a40aa79a22 100644 --- a/docs/mkdocs_github_authors.yaml +++ b/docs/mkdocs_github_authors.yaml @@ -4,6 +4,9 @@ 130829914+IvorZhu331@users.noreply.github.com: avatar: https://avatars.githubusercontent.com/u/130829914?v=4 username: IvorZhu331 +131261051+MatthewNoyce@users.noreply.github.com: + avatar: https://avatars.githubusercontent.com/u/131261051?v=4 + username: MatthewNoyce 135830346+UltralyticsAssistant@users.noreply.github.com: avatar: https://avatars.githubusercontent.com/u/135830346?v=4 username: UltralyticsAssistant @@ -97,6 +100,9 @@ lakshantha@ultralytics.com: lakshanthad@yahoo.com: avatar: https://avatars.githubusercontent.com/u/20147381?v=4 username: lakshanthad +matthewnoyce@icloud.com: + avatar: https://avatars.githubusercontent.com/u/131261051?v=4 + username: MatthewNoyce muhammadrizwanmunawar123@gmail.com: avatar: https://avatars.githubusercontent.com/u/62513924?v=4 username: RizwanMunawar