diff --git a/.github/ISSUE_TEMPLATE/bug-report.yml b/.github/ISSUE_TEMPLATE/bug-report.yml
index 430b05957a..0631213715 100644
--- a/.github/ISSUE_TEMPLATE/bug-report.yml
+++ b/.github/ISSUE_TEMPLATE/bug-report.yml
@@ -56,7 +56,7 @@ body:
placeholder: |
Paste output of `yolo checks` or `ultralytics.checks()` command, i.e.:
```
- Ultralytics YOLOv8.0.181 🚀 Python-3.11.2 torch-2.0.1 CPU (Apple M2)
+ Ultralytics 8.3.2 🚀 Python-3.11.2 torch-2.4.1 CPU (Apple M3)
Setup complete ✅ (8 CPUs, 16.0 GB RAM, 266.5/460.4 GB disk)
OS macOS-13.5.2
@@ -64,7 +64,7 @@ body:
Python 3.11.2
Install git
RAM 16.00 GB
- CPU Apple M2
+ CPU Apple M3
CUDA None
```
validations:
diff --git a/.github/ISSUE_TEMPLATE/config.yml b/.github/ISSUE_TEMPLATE/config.yml
index 73745a3a56..99cadfd893 100644
--- a/.github/ISSUE_TEMPLATE/config.yml
+++ b/.github/ISSUE_TEMPLATE/config.yml
@@ -4,7 +4,7 @@ blank_issues_enabled: true
contact_links:
- name: 📄 Docs
url: https://docs.ultralytics.com/
- about: Full Ultralytics YOLOv8 Documentation
+ about: Full Ultralytics YOLO Documentation
- name: 💬 Forum
url: https://community.ultralytics.com/
about: Ask on Ultralytics Community Forum
diff --git a/.github/ISSUE_TEMPLATE/feature-request.yml b/.github/ISSUE_TEMPLATE/feature-request.yml
index c065446c1f..04cefb18cf 100644
--- a/.github/ISSUE_TEMPLATE/feature-request.yml
+++ b/.github/ISSUE_TEMPLATE/feature-request.yml
@@ -1,14 +1,14 @@
# Ultralytics YOLO 🚀, AGPL-3.0 license
name: 🚀 Feature Request
-description: Suggest a YOLOv8 idea
+description: Suggest an Ultralytics YOLO idea
# title: " "
labels: [enhancement]
body:
- type: markdown
attributes:
value: |
- Thank you for submitting a YOLOv8 🚀 Feature Request!
+ Thank you for submitting an Ultralytics 🚀 Feature Request!
- type: checkboxes
attributes:
@@ -17,7 +17,7 @@ body:
Please search the Ultralytics [Docs](https://docs.ultralytics.com) and [issues](https://github.com/ultralytics/ultralytics/issues) to see if a similar feature request already exists.
options:
- label: >
- I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar feature requests.
+ I have searched the Ultralytics [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar feature requests.
required: true
- type: textarea
@@ -25,7 +25,7 @@ body:
label: Description
description: A short description of your feature.
placeholder: |
- What new feature would you like to see in YOLOv8?
+ What new feature would you like to see in YOLO?
validations:
required: true
@@ -46,7 +46,7 @@ body:
attributes:
label: Are you willing to submit a PR?
description: >
- (Optional) We encourage you to submit a [Pull Request](https://github.com/ultralytics/ultralytics/pulls) (PR) to help improve YOLOv8 for everyone, especially if you have a good understanding of how to implement a fix or feature.
- See the YOLOv8 [Contributing Guide](https://docs.ultralytics.com/help/contributing) to get started.
+ (Optional) We encourage you to submit a [Pull Request](https://github.com/ultralytics/ultralytics/pulls) (PR) to help improve YOLO for everyone, especially if you have a good understanding of how to implement a fix or feature.
+ See the Ultralytics [Contributing Guide](https://docs.ultralytics.com/help/contributing) to get started.
options:
- label: Yes I'd like to help by submitting a PR!
diff --git a/.github/workflows/ci.yaml b/.github/workflows/ci.yaml
index d98402e712..c54ee4494c 100644
--- a/.github/workflows/ci.yaml
+++ b/.github/workflows/ci.yaml
@@ -9,7 +9,7 @@ on:
pull_request:
branches: [main]
schedule:
- - cron: "0 0 * * *" # runs at 00:00 UTC every day
+ - cron: "0 8 * * *" # runs at 08:00 UTC every day
workflow_dispatch:
inputs:
hub:
@@ -256,7 +256,7 @@ jobs:
- name: Benchmark ClassificationModel
run: python -m ultralytics.cfg.__init__ benchmark model='yolo11n-cls.pt' imgsz=160 verbose=0.249
- name: Benchmark YOLOWorld DetectionModel
- run: python -m ultralytics.cfg.__init__ benchmark model='yolo11s-worldv2.pt' imgsz=160 verbose=0.337
+ run: python -m ultralytics.cfg.__init__ benchmark model='yolov8s-worldv2.pt' imgsz=160 verbose=0.337
- name: Benchmark SegmentationModel
run: python -m ultralytics.cfg.__init__ benchmark model='yolo11n-seg.pt' imgsz=160 verbose=0.195
- name: Benchmark PoseModel
@@ -320,16 +320,16 @@ jobs:
conda list
- name: Test CLI
run: |
- yolo predict model=yolov8n.pt imgsz=320
- yolo train model=yolov8n.pt data=coco8.yaml epochs=1 imgsz=32
- yolo val model=yolov8n.pt data=coco8.yaml imgsz=32
- yolo export model=yolov8n.pt format=torchscript imgsz=160
+ yolo predict model=yolo11n.pt imgsz=320
+ yolo train model=yolo11n.pt data=coco8.yaml epochs=1 imgsz=32
+ yolo val model=yolo11n.pt data=coco8.yaml imgsz=32
+ yolo export model=yolo11n.pt format=torchscript imgsz=160
- name: Test Python
# Note this step must use the updated default bash environment, not a python environment
run: |
python -c "
from ultralytics import YOLO
- model = YOLO('yolov8n.pt')
+ model = YOLO('yolo11n.pt')
results = model.train(data='coco8.yaml', epochs=3, imgsz=160)
results = model.val(imgsz=160)
results = model.predict(imgsz=160)
diff --git a/.github/workflows/format.yml b/.github/workflows/format.yml
index 9befe5c968..516450f876 100644
--- a/.github/workflows/format.yml
+++ b/.github/workflows/format.yml
@@ -48,7 +48,7 @@ jobs:
## Environments
- YOLOv8 may be run in any of the following up-to-date verified environments (with all dependencies including [CUDA](https://developer.nvidia.com/cuda)/[CUDNN](https://developer.nvidia.com/cudnn), [Python](https://www.python.org/) and [PyTorch](https://pytorch.org/) preinstalled):
+ YOLO may be run in any of the following up-to-date verified environments (with all dependencies including [CUDA](https://developer.nvidia.com/cuda)/[CUDNN](https://developer.nvidia.com/cudnn), [Python](https://www.python.org/) and [PyTorch](https://pytorch.org/) preinstalled):
- **Notebooks** with free GPU:
- **Google Cloud** Deep Learning VM. See [GCP Quickstart Guide](https://docs.ultralytics.com/yolov5/environments/google_cloud_quickstart_tutorial/)
@@ -59,4 +59,4 @@ jobs:
- If this badge is green, all [Ultralytics CI](https://github.com/ultralytics/ultralytics/actions/workflows/ci.yaml?query=event%3Aschedule) tests are currently passing. CI tests verify correct operation of all YOLOv8 [Modes](https://docs.ultralytics.com/modes/) and [Tasks](https://docs.ultralytics.com/tasks/) on macOS, Windows, and Ubuntu every 24 hours and on every commit.
+ If this badge is green, all [Ultralytics CI](https://github.com/ultralytics/ultralytics/actions/workflows/ci.yaml?query=event%3Aschedule) tests are currently passing. CI tests verify correct operation of all YOLO [Modes](https://docs.ultralytics.com/modes/) and [Tasks](https://docs.ultralytics.com/tasks/) on macOS, Windows, and Ubuntu every 24 hours and on every commit.
diff --git a/README.md b/README.md
index 851d6a2d7b..b39b089a2e 100644
--- a/README.md
+++ b/README.md
@@ -126,11 +126,11 @@ See [Detection Docs](https://docs.ultralytics.com/tasks/detect/) for usage examp
| Model | size
(pixels) | mAPval
50-95 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) |
| ------------------------------------------------------------------------------------ | --------------------- | -------------------- | ------------------------------ | ----------------------------------- | ------------------ | ----------------- |
-| [YOLO11n](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n.pt) | 640 | 39.5 | 56.12 ± 0.82 ms | 1.55 ± 0.01 ms | 2.6 | 6.5 |
-| [YOLO11s](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s.pt) | 640 | 47.0 | 90.01 ± 1.17 ms | 2.46 ± 0.00 ms | 9.4 | 21.5 |
-| [YOLO11m](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m.pt) | 640 | 51.5 | 183.20 ± 2.04 ms | 4.70 ± 0.06 ms | 20.1 | 68.0 |
-| [YOLO11l](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l.pt) | 640 | 53.4 | 238.64 ± 1.39 ms | 6.16 ± 0.08 ms | 25.3 | 86.9 |
-| [YOLO11x](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x.pt) | 640 | 54.7 | 462.78 ± 6.66 ms | 11.31 ± 0.24 ms | 56.9 | 194.9 |
+| [YOLO11n](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n.pt) | 640 | 39.5 | 56.1 ± 0.8 | 1.5 ± 0.0 | 2.6 | 6.5 |
+| [YOLO11s](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s.pt) | 640 | 47.0 | 90.0 ± 1.2 | 2.5 ± 0.0 | 9.4 | 21.5 |
+| [YOLO11m](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m.pt) | 640 | 51.5 | 183.2 ± 2.0 | 4.7 ± 0.1 | 20.1 | 68.0 |
+| [YOLO11l](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l.pt) | 640 | 53.4 | 238.6 ± 1.4 | 6.2 ± 0.1 | 25.3 | 86.9 |
+| [YOLO11x](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x.pt) | 640 | 54.7 | 462.8 ± 6.7 | 11.3 ± 0.2 | 56.9 | 194.9 |
- **mAPval** values are for single-model single-scale on [COCO val2017](https://cocodataset.org/) dataset.
Reproduce by `yolo val detect data=coco.yaml device=0`
- **Speed** averaged over COCO val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val detect data=coco.yaml batch=1 device=0|cpu`
@@ -143,28 +143,45 @@ See [Segmentation Docs](https://docs.ultralytics.com/tasks/segment/) for usage e
| Model | size
(pixels) | mAPbox
50-95 | mAPmask
50-95 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) |
| -------------------------------------------------------------------------------------------- | --------------------- | -------------------- | --------------------- | ------------------------------ | ----------------------------------- | ------------------ | ----------------- |
-| [YOLO11n-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-seg.pt) | 640 | 38.9 | 32.0 | 65.90 ± 1.14 ms | 1.84 ± 0.00 ms | 2.9 | 10.4 |
-| [YOLO11s-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-seg.pt) | 640 | 46.6 | 37.8 | 117.56 ± 4.89 ms | 2.94 ± 0.01 ms | 10.1 | 35.5 |
-| [YOLO11m-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-seg.pt) | 640 | 51.5 | 41.5 | 281.63 ± 1.16 ms | 6.31 ± 0.09 ms | 22.4 | 123.3 |
-| [YOLO11l-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-seg.pt) | 640 | 53.4 | 42.9 | 344.16 ± 3.17 ms | 7.78 ± 0.16 ms | 27.6 | 142.2 |
-| [YOLO11x-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-seg.pt) | 640 | 54.7 | 43.8 | 664.50 ± 3.24 ms | 15.75 ± 0.67 ms | 62.1 | 319.0 |
+| [YOLO11n-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-seg.pt) | 640 | 38.9 | 32.0 | 65.9 ± 1.1 | 1.8 ± 0.0 | 2.9 | 10.4 |
+| [YOLO11s-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-seg.pt) | 640 | 46.6 | 37.8 | 117.6 ± 4.9 | 2.9 ± 0.0 | 10.1 | 35.5 |
+| [YOLO11m-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-seg.pt) | 640 | 51.5 | 41.5 | 281.6 ± 1.2 | 6.3 ± 0.1 | 22.4 | 123.3 |
+| [YOLO11l-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-seg.pt) | 640 | 53.4 | 42.9 | 344.2 ± 3.2 | 7.8 ± 0.2 | 27.6 | 142.2 |
+| [YOLO11x-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-seg.pt) | 640 | 54.7 | 43.8 | 664.5 ± 3.2 | 15.8 ± 0.7 | 62.1 | 319.0 |
- **mAPval** values are for single-model single-scale on [COCO val2017](https://cocodataset.org/) dataset.
Reproduce by `yolo val segment data=coco-seg.yaml device=0`
- **Speed** averaged over COCO val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val segment data=coco-seg.yaml batch=1 device=0|cpu`
+Classification (ImageNet)
+
+See [Classification Docs](https://docs.ultralytics.com/tasks/classify/) for usage examples with these models trained on [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/), which include 1000 pretrained classes.
+
+| Model | size
(pixels) | acc
top1 | acc
top5 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) at 640 |
+| -------------------------------------------------------------------------------------------- | --------------------- | ---------------- | ---------------- | ------------------------------ | ----------------------------------- | ------------------ | ------------------------ |
+| [YOLO11n-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-cls.pt) | 224 | 70.0 | 89.4 | 5.0 ± 0.3 | 1.1 ± 0.0 | 1.6 | 3.3 |
+| [YOLO11s-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-cls.pt) | 224 | 75.4 | 92.7 | 7.9 ± 0.2 | 1.3 ± 0.0 | 5.5 | 12.1 |
+| [YOLO11m-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-cls.pt) | 224 | 77.3 | 93.9 | 17.2 ± 0.4 | 2.0 ± 0.0 | 10.4 | 39.3 |
+| [YOLO11l-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-cls.pt) | 224 | 78.3 | 94.3 | 23.2 ± 0.3 | 2.8 ± 0.0 | 12.9 | 49.4 |
+| [YOLO11x-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-cls.pt) | 224 | 79.5 | 94.9 | 41.4 ± 0.9 | 3.8 ± 0.0 | 28.4 | 110.4 |
+
+- **acc** values are model accuracies on the [ImageNet](https://www.image-net.org/) dataset validation set.
Reproduce by `yolo val classify data=path/to/ImageNet device=0`
+- **Speed** averaged over ImageNet val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val classify data=path/to/ImageNet batch=1 device=0|cpu`
+
+Pose (COCO)
See [Pose Docs](https://docs.ultralytics.com/tasks/pose/) for usage examples with these models trained on [COCO-Pose](https://docs.ultralytics.com/datasets/pose/coco/), which include 1 pre-trained class, person.
| Model | size
(pixels) | mAPpose
50-95 | mAPpose
50 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) |
| ---------------------------------------------------------------------------------------------- | --------------------- | --------------------- | ------------------ | ------------------------------ | ----------------------------------- | ------------------ | ----------------- |
-| [YOLO11n-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-pose.pt) | 640 | 50.0 | 81.0 | 52.40 ± 0.51 ms | 1.72 ± 0.01 ms | 2.9 | 7.6 |
-| [YOLO11s-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-pose.pt) | 640 | 58.9 | 86.3 | 90.54 ± 0.59 ms | 2.57 ± 0.00 ms | 9.9 | 23.2 |
-| [YOLO11m-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-pose.pt) | 640 | 64.9 | 89.4 | 187.28 ± 0.77 ms | 4.94 ± 0.05 ms | 20.9 | 71.7 |
-| [YOLO11l-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-pose.pt) | 640 | 66.1 | 89.9 | 247.69 ± 1.10 ms | 6.42 ± 0.13 ms | 26.2 | 90.7 |
-| [YOLO11x-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-pose.pt) | 640 | 69.5 | 91.1 | 487.97 ± 13.91 ms | 12.06 ± 0.20 ms | 58.8 | 203.3 |
+| [YOLO11n-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-pose.pt) | 640 | 50.0 | 81.0 | 52.4 ± 0.5 | 1.7 ± 0.0 | 2.9 | 7.6 |
+| [YOLO11s-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-pose.pt) | 640 | 58.9 | 86.3 | 90.5 ± 0.6 | 2.6 ± 0.0 | 9.9 | 23.2 |
+| [YOLO11m-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-pose.pt) | 640 | 64.9 | 89.4 | 187.3 ± 0.8 | 4.9 ± 0.1 | 20.9 | 71.7 |
+| [YOLO11l-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-pose.pt) | 640 | 66.1 | 89.9 | 247.7 ± 1.1 | 6.4 ± 0.1 | 26.2 | 90.7 |
+| [YOLO11x-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-pose.pt) | 640 | 69.5 | 91.1 | 488.0 ± 13.9 | 12.1 ± 0.2 | 58.8 | 203.3 |
- **mAPval** values are for single-model single-scale on [COCO Keypoints val2017](https://cocodataset.org/) dataset.
Reproduce by `yolo val pose data=coco-pose.yaml device=0`
- **Speed** averaged over COCO val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val pose data=coco-pose.yaml batch=1 device=0|cpu`
@@ -177,34 +194,17 @@ See [OBB Docs](https://docs.ultralytics.com/tasks/obb/) for usage examples with
| Model | size
(pixels) | mAPtest
50 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) |
| -------------------------------------------------------------------------------------------- | --------------------- | ------------------ | ------------------------------ | ----------------------------------- | ------------------ | ----------------- |
-| [YOLO11n-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-obb.pt) | 1024 | 78.4 | 117.56 ± 0.80 ms | 4.43 ± 0.01 ms | 2.7 | 17.2 |
-| [YOLO11s-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-obb.pt) | 1024 | 79.5 | 219.41 ± 4.00 ms | 5.13 ± 0.02 ms | 9.7 | 57.5 |
-| [YOLO11m-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-obb.pt) | 1024 | 80.9 | 562.81 ± 2.87 ms | 10.07 ± 0.38 ms | 20.9 | 183.5 |
-| [YOLO11l-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-obb.pt) | 1024 | 81.0 | 712.49 ± 4.98 ms | 13.46 ± 0.55 ms | 26.2 | 232.0 |
-| [YOLO11x-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-obb.pt) | 1024 | 81.3 | 1408.63 ± 7.67 ms | 28.59 ± 0.96 ms | 58.8 | 520.2 |
+| [YOLO11n-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-obb.pt) | 1024 | 78.4 | 117.6 ± 0.8 | 4.4 ± 0.0 | 2.7 | 17.2 |
+| [YOLO11s-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-obb.pt) | 1024 | 79.5 | 219.4 ± 4.0 | 5.1 ± 0.0 | 9.7 | 57.5 |
+| [YOLO11m-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-obb.pt) | 1024 | 80.9 | 562.8 ± 2.9 | 10.1 ± 0.4 | 20.9 | 183.5 |
+| [YOLO11l-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-obb.pt) | 1024 | 81.0 | 712.5 ± 5.0 | 13.5 ± 0.6 | 26.2 | 232.0 |
+| [YOLO11x-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-obb.pt) | 1024 | 81.3 | 1408.6 ± 7.7 | 28.6 ± 1.0 | 58.8 | 520.2 |
- **mAPtest** values are for single-model multiscale on [DOTAv1](https://captain-whu.github.io/DOTA/index.html) dataset.
Reproduce by `yolo val obb data=DOTAv1.yaml device=0 split=test` and submit merged results to [DOTA evaluation](https://captain-whu.github.io/DOTA/evaluation.html).
- **Speed** averaged over DOTAv1 val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val obb data=DOTAv1.yaml batch=1 device=0|cpu`
Classification (ImageNet)
-
-See [Classification Docs](https://docs.ultralytics.com/tasks/classify/) for usage examples with these models trained on [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/), which include 1000 pretrained classes.
-
-| Model | size
(pixels) | acc
top1 | acc
top5 | Speed
CPU ONNX
(ms) | Speed
T4 TensorRT10
(ms) | params
(M) | FLOPs
(B) at 640 |
-| -------------------------------------------------------------------------------------------- | --------------------- | ---------------- | ---------------- | ------------------------------ | ----------------------------------- | ------------------ | ------------------------ |
-| [YOLO11n-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-cls.pt) | 224 | 70.0 | 89.4 | 5.03 ± 0.32 ms | 1.10 ± 0.01 ms | 1.6 | 3.3 |
-| [YOLO11s-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-cls.pt) | 224 | 75.4 | 92.7 | 7.89 ± 0.18 ms | 1.34 ± 0.01 ms | 5.5 | 12.1 |
-| [YOLO11m-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-cls.pt) | 224 | 77.3 | 93.9 | 17.17 ± 0.40 ms | 1.95 ± 0.00 ms | 10.4 | 39.3 |
-| [YOLO11l-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-cls.pt) | 224 | 78.3 | 94.3 | 23.17 ± 0.29 ms | 2.76 ± 0.00 ms | 12.9 | 49.4 |
-| [YOLO11x-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-cls.pt) | 224 | 79.5 | 94.9 | 41.41 ± 0.94 ms | 3.82 ± 0.00 ms | 28.4 | 110.4 |
-
-- **acc** values are model accuracies on the [ImageNet](https://www.image-net.org/) dataset validation set.
Reproduce by `yolo val classify data=path/to/ImageNet device=0`
-- **Speed** averaged over ImageNet val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val classify data=path/to/ImageNet batch=1 device=0|cpu`
-
-
(像素) | mAPval
50-95 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) |
| ------------------------------------------------------------------------------------ | ------------------- | -------------------- | ----------------------------- | ---------------------------------- | ---------------- | ----------------- |
-| [YOLO11n](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n.pt) | 640 | 39.5 | 56.12 ± 0.82 ms | 1.55 ± 0.01 ms | 2.6 | 6.5 |
-| [YOLO11s](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s.pt) | 640 | 47.0 | 90.01 ± 1.17 ms | 2.46 ± 0.00 ms | 9.4 | 21.5 |
-| [YOLO11m](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m.pt) | 640 | 51.5 | 183.20 ± 2.04 ms | 4.70 ± 0.06 ms | 20.1 | 68.0 |
-| [YOLO11l](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l.pt) | 640 | 53.4 | 238.64 ± 1.39 ms | 6.16 ± 0.08 ms | 25.3 | 86.9 |
-| [YOLO11x](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x.pt) | 640 | 54.7 | 462.78 ± 6.66 ms | 11.31 ± 0.24 ms | 56.9 | 194.9 |
+| [YOLO11n](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n.pt) | 640 | 39.5 | 56.1 ± 0.8 | 1.5 ± 0.0 | 2.6 | 6.5 |
+| [YOLO11s](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s.pt) | 640 | 47.0 | 90.0 ± 1.2 | 2.5 ± 0.0 | 9.4 | 21.5 |
+| [YOLO11m](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m.pt) | 640 | 51.5 | 183.2 ± 2.0 | 4.7 ± 0.1 | 20.1 | 68.0 |
+| [YOLO11l](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l.pt) | 640 | 53.4 | 238.6 ± 1.4 | 6.2 ± 0.1 | 25.3 | 86.9 |
+| [YOLO11x](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x.pt) | 640 | 54.7 | 462.8 ± 6.7 | 11.3 ± 0.2 | 56.9 | 194.9 |
- **mAPval** 值针对单模型单尺度在 [COCO val2017](https://cocodataset.org/) 数据集上进行。
复制命令 `yolo val detect data=coco.yaml device=0`
- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 COCO 验证图像上平均。
复制命令 `yolo val detect data=coco.yaml batch=1 device=0|cpu`
@@ -143,28 +143,45 @@ YOLO11 [检测](https://docs.ultralytics.com/tasks/detect/)、[分割](https://d
| 模型 | 尺寸
(像素) | mAPbox
50-95 | mAPmask
50-95 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) |
| -------------------------------------------------------------------------------------------- | ------------------- | -------------------- | --------------------- | ----------------------------- | ---------------------------------- | ---------------- | ----------------- |
-| [YOLO11n-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-seg.pt) | 640 | 38.9 | 32.0 | 65.90 ± 1.14 ms | 1.84 ± 0.00 ms | 2.9 | 10.4 |
-| [YOLO11s-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-seg.pt) | 640 | 46.6 | 37.8 | 117.56 ± 4.89 ms | 2.94 ± 0.01 ms | 10.1 | 35.5 |
-| [YOLO11m-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-seg.pt) | 640 | 51.5 | 41.5 | 281.63 ± 1.16 ms | 6.31 ± 0.09 ms | 22.4 | 123.3 |
-| [YOLO11l-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-seg.pt) | 640 | 53.4 | 42.9 | 344.16 ± 3.17 ms | 7.78 ± 0.16 ms | 27.6 | 142.2 |
-| [YOLO11x-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-seg.pt) | 640 | 54.7 | 43.8 | 664.50 ± 3.24 ms | 15.75 ± 0.67 ms | 62.1 | 319.0 |
+| [YOLO11n-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-seg.pt) | 640 | 38.9 | 32.0 | 65.9 ± 1.1 | 1.8 ± 0.0 | 2.9 | 10.4 |
+| [YOLO11s-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-seg.pt) | 640 | 46.6 | 37.8 | 117.6 ± 4.9 | 2.9 ± 0.0 | 10.1 | 35.5 |
+| [YOLO11m-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-seg.pt) | 640 | 51.5 | 41.5 | 281.6 ± 1.2 | 6.3 ± 0.1 | 22.4 | 123.3 |
+| [YOLO11l-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-seg.pt) | 640 | 53.4 | 42.9 | 344.2 ± 3.2 | 7.8 ± 0.2 | 27.6 | 142.2 |
+| [YOLO11x-seg](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-seg.pt) | 640 | 54.7 | 43.8 | 664.5 ± 3.2 | 15.8 ± 0.7 | 62.1 | 319.0 |
- **mAPval** 值针对单模型单尺度在 [COCO val2017](https://cocodataset.org/) 数据集上进行。
复制命令 `yolo val segment data=coco-seg.yaml device=0`
- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 COCO 验证图像上平均。
复制命令 `yolo val segment data=coco-seg.yaml batch=1 device=0|cpu`
+分类 (ImageNet)
+
+请参阅 [分类文档](https://docs.ultralytics.com/tasks/classify/) 以获取使用这些在 [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/) 数据集上训练的模型的示例,其中包含 1000 个预训练类别。
+
+| 模型 | 尺寸
(像素) | acc
top1 | acc
top5 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) at 640 |
+| -------------------------------------------------------------------------------------------- | ------------------- | ---------------- | ---------------- | ----------------------------- | ---------------------------------- | ---------------- | ------------------------ |
+| [YOLO11n-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-cls.pt) | 224 | 70.0 | 89.4 | 5.0 ± 0.3 | 1.1 ± 0.0 | 1.6 | 3.3 |
+| [YOLO11s-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-cls.pt) | 224 | 75.4 | 92.7 | 7.9 ± 0.2 | 1.3 ± 0.0 | 5.5 | 12.1 |
+| [YOLO11m-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-cls.pt) | 224 | 77.3 | 93.9 | 17.2 ± 0.4 | 2.0 ± 0.0 | 10.4 | 39.3 |
+| [YOLO11l-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-cls.pt) | 224 | 78.3 | 94.3 | 23.2 ± 0.3 | 2.8 ± 0.0 | 12.9 | 49.4 |
+| [YOLO11x-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-cls.pt) | 224 | 79.5 | 94.9 | 41.4 ± 0.9 | 3.8 ± 0.0 | 28.4 | 110.4 |
+
+- **acc** 值为在 [ImageNet](https://www.image-net.org/) 数据集验证集上的模型准确率。
复制命令 `yolo val classify data=path/to/ImageNet device=0`
+- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 ImageNet 验证图像上平均。
复制命令 `yolo val classify data=path/to/ImageNet batch=1 device=0|cpu`
+
+姿态 (COCO)
请参阅 [姿态文档](https://docs.ultralytics.com/tasks/pose/) 以获取使用这些在 [COCO-Pose](https://docs.ultralytics.com/datasets/pose/coco/) 数据集上训练的模型的示例,其中包含 1 个预训练类别(人)。
-| 模型 | 尺寸
(像素) | mAPpose
50-95 | mAPpose
50 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) |
-| ---------------------------------------------------------------------------------------------- | ------------------- | --------------------- | ------------------ | ----------------------------- | ---------------------------------- | ---------------- | ----------------- |
-| [YOLO11n-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-pose.pt) | 640 | 50.0 | 81.0 | 52.40 ± 0.51 ms | 1.72 ± 0.01 ms | 2.9 | 7.6 |
-| [YOLO11s-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-pose.pt) | 640 | 58.9 | 86.3 | 90.54 ± 0.59 ms | 2.57 ± 0.00 ms | 9.9 | 23.2 |
-| [YOLO11m-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-pose.pt) | 640 | 64.9 | 89.4 | 187.28 ± 0.77 ms | 4.94 ± 0.05 ms | 20.9 | 71.7 |
-| [YOLO11l-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-pose.pt) | 640 | 66.1 | 89.9 | 247.69 ± 1.10 ms | 6.42 ± 0.13 ms | 26.2 | 90.7 |
-| [YOLO11x-pose](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-pose.pt) | 640 | 69.5 | 91.1 | 487.97 ± 13.91 ms | 12.06 ± 0.20 ms | 58.8 | 203.3 |
+| 模型 | 尺寸
(像素) | mAPpose
50-95 | mAPpose
50 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) |
+| -------------------------------------------------------------------------------------------- | ------------------- | --------------------- | ------------------ | ----------------------------- | ---------------------------------- | ---------------- | ----------------- |
+| [YOLO11n-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-obb.pt) | 1024 | 78.4 | 117.6 ± 0.8 | 4.4 ± 0.0 | 2.7 | 17.2 |
+| [YOLO11s-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-obb.pt) | 1024 | 79.5 | 219.4 ± 4.0 | 5.1 ± 0.0 | 9.7 | 57.5 |
+| [YOLO11m-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-obb.pt) | 1024 | 80.9 | 562.8 ± 2.9 | 10.1 ± 0.4 | 20.9 | 183.5 |
+| [YOLO11l-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-obb.pt) | 1024 | 81.0 | 712.5 ± 5.0 | 13.5 ± 0.6 | 26.2 | 232.0 |
+| [YOLO11x-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-obb.pt) | 1024 | 81.3 | 1408.6 ± 7.7 | 28.6 ± 1.0 | 58.8 | 520.2 |
- **mAPval** 值针对单模型单尺度在 [COCO Keypoints val2017](https://cocodataset.org/) 数据集上进行。
复制命令 `yolo val pose data=coco-pose.yaml device=0`
- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 COCO 验证图像上平均。
复制命令 `yolo val pose data=coco-pose.yaml batch=1 device=0|cpu`
@@ -177,34 +194,17 @@ YOLO11 [检测](https://docs.ultralytics.com/tasks/detect/)、[分割](https://d
| 模型 | 尺寸
(像素) | mAPtest
50 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) |
| -------------------------------------------------------------------------------------------- | ------------------- | ------------------ | ----------------------------- | ---------------------------------- | ---------------- | ----------------- |
-| [YOLO11n-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-obb.pt) | 1024 | 78.4 | 117.56 ± 0.80 ms | 4.43 ± 0.01 ms | 2.7 | 17.2 |
-| [YOLO11s-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-obb.pt) | 1024 | 79.5 | 219.41 ± 4.00 ms | 5.13 ± 0.02 ms | 9.7 | 57.5 |
-| [YOLO11m-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-obb.pt) | 1024 | 80.9 | 562.81 ± 2.87 ms | 10.07 ± 0.38 ms | 20.9 | 183.5 |
-| [YOLO11l-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-obb.pt) | 1024 | 81.0 | 712.49 ± 4.98 ms | 13.46 ± 0.55 ms | 26.2 | 232.0 |
-| [YOLO11x-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-obb.pt) | 1024 | 81.3 | 1408.63 ± 7.67 ms | 28.59 ± 0.96 ms | 58.8 | 520.2 |
+| [YOLO11n-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-obb.pt) | 1024 | 78.4 | 117.56 ± 0.80 | 4.43 ± 0.01 | 2.7 | 17.2 |
+| [YOLO11s-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-obb.pt) | 1024 | 79.5 | 219.41 ± 4.00 | 5.13 ± 0.02 | 9.7 | 57.5 |
+| [YOLO11m-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-obb.pt) | 1024 | 80.9 | 562.81 ± 2.87 | 10.07 ± 0.38 | 20.9 | 183.5 |
+| [YOLO11l-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-obb.pt) | 1024 | 81.0 | 712.49 ± 4.98 | 13.46 ± 0.55 | 26.2 | 232.0 |
+| [YOLO11x-obb](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-obb.pt) | 1024 | 81.3 | 1408.63 ± 7.67 | 28.59 ± 0.96 | 58.8 | 520.2 |
- **mAPtest** 值针对单模型多尺度在 [DOTAv1](https://captain-whu.github.io/DOTA/index.html) 数据集上进行。
复制命令 `yolo val obb data=DOTAv1.yaml device=0 split=test` 并提交合并结果到 [DOTA 评估](https://captain-whu.github.io/DOTA/evaluation.html)。
- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 DOTAv1 验证图像上平均。
复制命令 `yolo val obb data=DOTAv1.yaml batch=1 device=0|cpu`
分类 (ImageNet)
-
-请参阅 [分类文档](https://docs.ultralytics.com/tasks/classify/) 以获取使用这些在 [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/) 数据集上训练的模型的示例,其中包含 1000 个预训练类别。
-
-| 模型 | 尺寸
(像素) | acc
top1 | acc
top5 | 速度
CPU ONNX
(ms) | 速度
T4 TensorRT10
(ms) | 参数
(M) | FLOPs
(B) at 640 |
-| -------------------------------------------------------------------------------------------- | ------------------- | ---------------- | ---------------- | ----------------------------- | ---------------------------------- | ---------------- | ------------------------ |
-| [YOLO11n-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11n-cls.pt) | 224 | 70.0 | 89.4 | 5.03 ± 0.32 ms | 1.10 ± 0.01 ms | 1.6 | 3.3 |
-| [YOLO11s-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11s-cls.pt) | 224 | 75.4 | 92.7 | 7.89 ± 0.18 ms | 1.34 ± 0.01 ms | 5.5 | 12.1 |
-| [YOLO11m-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11m-cls.pt) | 224 | 77.3 | 93.9 | 17.17 ± 0.40 ms | 1.95 ± 0.00 ms | 10.4 | 39.3 |
-| [YOLO11l-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11l-cls.pt) | 224 | 78.3 | 94.3 | 23.17 ± 0.29 ms | 2.76 ± 0.00 ms | 12.9 | 49.4 |
-| [YOLO11x-cls](https://github.com/ultralytics/assets/releases/download/v8.3.0/yolo11x-cls.pt) | 224 | 79.5 | 94.9 | 41.41 ± 0.94 ms | 3.82 ± 0.00 ms | 28.4 | 110.4 |
-
-- **acc** 值为在 [ImageNet](https://www.image-net.org/) 数据集验证集上的模型准确率。
复制命令 `yolo val classify data=path/to/ImageNet device=0`
-- **速度**在使用 [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) 实例的 ImageNet 验证图像上平均。
复制命令 `yolo val classify data=path/to/ImageNet batch=1 device=0|cpu`
-
-
- Watch: How to Train [Image Classification](https://www.ultralytics.com/glossary/image-classification) Model using Caltech-256 Dataset with Ultralytics HUB
+ Watch: How to Train Image Classification Model using Caltech-256 Dataset with Ultralytics HUB
@@ -22,7 +22,7 @@ This dataset is intended for use with [Ultralytics HUB](https://hub.ultralytics.
allowfullscreen>
- Watch: Train YOLOv8 Pose Model on Tiger-Pose Dataset Using Ultralytics HUB
+ Watch: Train YOLO11 Pose Model on Tiger-Pose Dataset Using Ultralytics HUB
@@ -152,9 +152,9 @@ Find comprehensive information on the [Predict](../modes/predict.md) page for fu ## FAQ -### What is a Coral Edge TPU and how does it enhance Raspberry Pi's performance with Ultralytics YOLOv8? +### What is a Coral Edge TPU and how does it enhance Raspberry Pi's performance with Ultralytics YOLO11? -The Coral Edge TPU is a compact device designed to add an Edge TPU coprocessor to your system. This coprocessor enables low-power, high-performance [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) inference, particularly optimized for TensorFlow Lite models. When using a Raspberry Pi, the Edge TPU accelerates ML model inference, significantly boosting performance, especially for Ultralytics YOLOv8 models. You can read more about the Coral Edge TPU on their [home page](https://coral.ai/products/accelerator). +The Coral Edge TPU is a compact device designed to add an Edge TPU coprocessor to your system. This coprocessor enables low-power, high-performance [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) inference, particularly optimized for TensorFlow Lite models. When using a Raspberry Pi, the Edge TPU accelerates ML model inference, significantly boosting performance, especially for Ultralytics YOLO11 models. You can read more about the Coral Edge TPU on their [home page](https://coral.ai/products/accelerator). ### How do I install the Coral Edge TPU runtime on a Raspberry Pi? @@ -166,9 +166,9 @@ sudo dpkg -i path/to/package.deb Make sure to uninstall any previous Coral Edge TPU runtime versions by following the steps outlined in the [Installation Walkthrough](#installation-walkthrough) section. -### Can I export my Ultralytics YOLOv8 model to be compatible with Coral Edge TPU? +### Can I export my Ultralytics YOLO11 model to be compatible with Coral Edge TPU? -Yes, you can export your Ultralytics YOLOv8 model to be compatible with the Coral Edge TPU. It is recommended to perform the export on Google Colab, an x86_64 Linux machine, or using the [Ultralytics Docker container](docker-quickstart.md). You can also use Ultralytics HUB for exporting. Here is how you can export your model using Python and CLI: +Yes, you can export your Ultralytics YOLO11 model to be compatible with the Coral Edge TPU. It is recommended to perform the export on Google Colab, an x86_64 Linux machine, or using the [Ultralytics Docker container](docker-quickstart.md). You can also use Ultralytics HUB for exporting. Here is how you can export your model using Python and CLI: !!! note "Exporting the model" @@ -208,9 +208,9 @@ pip install -U tflite-runtime For a specific wheel, such as TensorFlow 2.15.0 `tflite-runtime`, you can download it from [this link](https://github.com/feranick/TFlite-builds/releases) and install it using `pip`. Detailed instructions are available in the section on running the model [Running the Model](#running-the-model). -### How do I run inference with an exported YOLOv8 model on a Raspberry Pi using the Coral Edge TPU? +### How do I run inference with an exported YOLO11 model on a Raspberry Pi using the Coral Edge TPU? -After exporting your YOLOv8 model to an Edge TPU-compatible format, you can run inference using the following code snippets: +After exporting your YOLO11 model to an Edge TPU-compatible format, you can run inference using the following code snippets: !!! note "Running the model" diff --git a/docs/en/guides/data-collection-and-annotation.md b/docs/en/guides/data-collection-and-annotation.md index dce15e0a68..058323ee29 100644 --- a/docs/en/guides/data-collection-and-annotation.md +++ b/docs/en/guides/data-collection-and-annotation.md @@ -136,12 +136,12 @@ Bouncing your ideas and queries off other [computer vision](https://www.ultralyt ### Where to Find Help and Support -- **GitHub Issues:** Visit the YOLOv8 GitHub repository and use the [Issues tab](https://github.com/ultralytics/ultralytics/issues) to raise questions, report bugs, and suggest features. The community and maintainers are there to help with any issues you face. +- **GitHub Issues:** Visit the YOLO11 GitHub repository and use the [Issues tab](https://github.com/ultralytics/ultralytics/issues) to raise questions, report bugs, and suggest features. The community and maintainers are there to help with any issues you face. - **Ultralytics Discord Server:** Join the [Ultralytics Discord server](https://discord.com/invite/ultralytics) to connect with other users and developers, get support, share knowledge, and brainstorm ideas. ### Official Documentation -- **Ultralytics YOLOv8 Documentation:** Refer to the [official YOLOv8 documentation](./index.md) for thorough guides and valuable insights on numerous computer vision tasks and projects. +- **Ultralytics YOLO11 Documentation:** Refer to the [official YOLO11 documentation](./index.md) for thorough guides and valuable insights on numerous computer vision tasks and projects. ## Conclusion @@ -159,7 +159,7 @@ Ensuring high consistency and accuracy in data annotation involves establishing ### How many images do I need for training Ultralytics YOLO models? -For effective [transfer learning](https://www.ultralytics.com/glossary/transfer-learning) and object detection with Ultralytics YOLO models, start with a minimum of a few hundred annotated objects per class. If training for just one class, begin with at least 100 annotated images and train for approximately 100 [epochs](https://www.ultralytics.com/glossary/epoch). More complex tasks might require thousands of images per class to achieve high reliability and performance. Quality annotations are crucial, so ensure your data collection and annotation processes are rigorous and aligned with your project's specific goals. Explore detailed training strategies in the [YOLOv8 training guide](../modes/train.md). +For effective [transfer learning](https://www.ultralytics.com/glossary/transfer-learning) and object detection with Ultralytics YOLO models, start with a minimum of a few hundred annotated objects per class. If training for just one class, begin with at least 100 annotated images and train for approximately 100 [epochs](https://www.ultralytics.com/glossary/epoch). More complex tasks might require thousands of images per class to achieve high reliability and performance. Quality annotations are crucial, so ensure your data collection and annotation processes are rigorous and aligned with your project's specific goals. Explore detailed training strategies in the [YOLO11 training guide](../modes/train.md). ### What are some popular tools for data annotation? diff --git a/docs/en/guides/deepstream-nvidia-jetson.md b/docs/en/guides/deepstream-nvidia-jetson.md index ab15009b99..90c361cb92 100644 --- a/docs/en/guides/deepstream-nvidia-jetson.md +++ b/docs/en/guides/deepstream-nvidia-jetson.md @@ -1,10 +1,10 @@ --- comments: true -description: Learn how to deploy Ultralytics YOLOv8 on NVIDIA Jetson devices using TensorRT and DeepStream SDK. Explore performance benchmarks and maximize AI capabilities. -keywords: Ultralytics, YOLOv8, NVIDIA Jetson, JetPack, AI deployment, embedded systems, deep learning, TensorRT, DeepStream SDK, computer vision +description: Learn how to deploy Ultralytics YOLO11 on NVIDIA Jetson devices using TensorRT and DeepStream SDK. Explore performance benchmarks and maximize AI capabilities. +keywords: Ultralytics, YOLO11, NVIDIA Jetson, JetPack, AI deployment, embedded systems, deep learning, TensorRT, DeepStream SDK, computer vision --- -# Ultralytics YOLOv8 on NVIDIA Jetson using DeepStream SDK and TensorRT +# Ultralytics YOLO11 on NVIDIA Jetson using DeepStream SDK and TensorRT
@@ -14,10 +14,10 @@ keywords: Ultralytics, YOLOv8, NVIDIA Jetson, JetPack, AI deployment, embedded s
allowfullscreen>
- Watch: How to Run Multiple Streams with DeepStream SDK on Jetson Nano using Ultralytics YOLOv8
+ Watch: How to Run Multiple Streams with DeepStream SDK on Jetson Nano using Ultralytics YOLO11
- +
Primary users include traffic management authorities and law enforcement, while secondary stakeholders are highway planners and the public benefiting from safer roads. Key requirements involve evaluating budget, time, and personnel, as well as addressing technical needs like high-resolution cameras and real-time data processing. Additionally, regulatory constraints on privacy and [data security](https://www.ultralytics.com/glossary/data-security) must be considered. @@ -85,7 +85,7 @@ The most popular computer vision tasks include [image classification](https://ww -For a detailed explanation of various tasks, please take a look at the Ultralytics Docs page on [YOLOv8 Tasks](../tasks/index.md). +For a detailed explanation of various tasks, please take a look at the Ultralytics Docs page on [YOLO11 Tasks](../tasks/index.md). ### Can a Pre-trained Model Remember Classes It Knew Before Custom Training? @@ -114,12 +114,12 @@ Connecting with other computer vision enthusiasts can be incredibly helpful for ### Community Support Channels -- **GitHub Issues:** Head over to the YOLOv8 GitHub repository. You can use the [Issues tab](https://github.com/ultralytics/ultralytics/issues) to raise questions, report bugs, and suggest features. The community and maintainers can assist with specific problems you encounter. +- **GitHub Issues:** Head over to the YOLO11 GitHub repository. You can use the [Issues tab](https://github.com/ultralytics/ultralytics/issues) to raise questions, report bugs, and suggest features. The community and maintainers can assist with specific problems you encounter. - **Ultralytics Discord Server:** Become part of the [Ultralytics Discord server](https://discord.com/invite/ultralytics). Connect with fellow users and developers, seek support, exchange knowledge, and discuss ideas. ### Comprehensive Guides and Documentation -- **Ultralytics YOLOv8 Documentation:** Explore the [official YOLOv8 documentation](./index.md) for in-depth guides and valuable tips on various computer vision tasks and projects. +- **Ultralytics YOLO11 Documentation:** Explore the [official YOLO11 documentation](./index.md) for in-depth guides and valuable tips on various computer vision tasks and projects. ## Conclusion @@ -138,11 +138,11 @@ To define a clear problem statement for your Ultralytics computer vision project Providing a well-defined problem statement ensures that the project remains focused and aligned with your objectives. For a detailed guide, refer to our [practical guide](#defining-a-clear-problem-statement). -### Why should I use Ultralytics YOLOv8 for speed estimation in my computer vision project? +### Why should I use Ultralytics YOLO11 for speed estimation in my computer vision project? -Ultralytics YOLOv8 is ideal for speed estimation because of its real-time object tracking capabilities, high accuracy, and robust performance in detecting and monitoring vehicle speeds. It overcomes inefficiencies and inaccuracies of traditional radar systems by leveraging cutting-edge computer vision technology. Check out our blog on [speed estimation using YOLOv8](https://www.ultralytics.com/blog/ultralytics-yolov8-for-speed-estimation-in-computer-vision-projects) for more insights and practical examples. +Ultralytics YOLO11 is ideal for speed estimation because of its real-time object tracking capabilities, high accuracy, and robust performance in detecting and monitoring vehicle speeds. It overcomes inefficiencies and inaccuracies of traditional radar systems by leveraging cutting-edge computer vision technology. Check out our blog on [speed estimation using YOLO11](https://www.ultralytics.com/blog/ultralytics-yolov8-for-speed-estimation-in-computer-vision-projects) for more insights and practical examples. -### How do I set effective measurable objectives for my computer vision project with Ultralytics YOLOv8? +### How do I set effective measurable objectives for my computer vision project with Ultralytics YOLO11? Set effective and measurable objectives using the SMART criteria: diff --git a/docs/en/guides/distance-calculation.md b/docs/en/guides/distance-calculation.md index 443b208b70..b0b12f919b 100644 --- a/docs/en/guides/distance-calculation.md +++ b/docs/en/guides/distance-calculation.md @@ -1,14 +1,14 @@ --- comments: true -description: Learn how to calculate distances between objects using Ultralytics YOLOv8 for accurate spatial positioning and scene understanding. -keywords: Ultralytics, YOLOv8, distance calculation, computer vision, object tracking, spatial positioning +description: Learn how to calculate distances between objects using Ultralytics YOLO11 for accurate spatial positioning and scene understanding. +keywords: Ultralytics, YOLO11, distance calculation, computer vision, object tracking, spatial positioning --- -# Distance Calculation using Ultralytics YOLOv8 +# Distance Calculation using Ultralytics YOLO11 ## What is Distance Calculation? -Measuring the gap between two objects is known as distance calculation within a specified space. In the case of [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics), the [bounding box](https://www.ultralytics.com/glossary/bounding-box) centroid is employed to calculate the distance for bounding boxes highlighted by the user. +Measuring the gap between two objects is known as distance calculation within a specified space. In the case of [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics), the [bounding box](https://www.ultralytics.com/glossary/bounding-box) centroid is employed to calculate the distance for bounding boxes highlighted by the user.
@@ -18,14 +18,14 @@ Measuring the gap between two objects is known as distance calculation within a
allowfullscreen>
- Watch: Distance Calculation using Ultralytics YOLOv8
+ Watch: Distance Calculation using Ultralytics YOLO11
@@ -18,7 +18,7 @@ A heatmap generated with [Ultralytics YOLOv8](https://github.com/ultralytics/ult
allowfullscreen>
- Watch: Heatmaps using Ultralytics YOLOv8
+ Watch: Heatmaps using Ultralytics YOLO11
@@ -18,16 +18,16 @@ Object blurring with [Ultralytics YOLOv8](https://github.com/ultralytics/ultraly
allowfullscreen>
- Watch: Object Blurring using Ultralytics YOLOv8
+ Watch: Object Blurring using Ultralytics YOLO11
- Watch: Class-wise Object Counting using Ultralytics YOLOv8 + Watch: Class-wise Object Counting using Ultralytics YOLO11 |
@@ -18,23 +18,23 @@ Object cropping with [Ultralytics YOLOv8](https://github.com/ultralytics/ultraly
allowfullscreen>
- Watch: Object Cropping using Ultralytics YOLOv8
+ Watch: Object Cropping using Ultralytics YOLO
@@ -18,21 +18,21 @@ Parking management with [Ultralytics YOLOv8](https://github.com/ultralytics/ultr
allowfullscreen>
- Watch: How to Implement Parking Management Using Ultralytics YOLOv8 🚀
+ Watch: How to Implement Parking Management Using Ultralytics YOLO 🚀
@@ -18,7 +18,7 @@ Queue management using [Ultralytics YOLOv8](https://github.com/ultralytics/ultra
allowfullscreen>
- Watch: How to Implement Queue Management with Ultralytics YOLOv8 | Airport and Metro Station
+ Watch: How to Implement Queue Management with Ultralytics YOLO11 | Airport and Metro Station
@@ -24,7 +24,7 @@ SAHI (Slicing Aided Hyper Inference) is an innovative library designed to optimi
allowfullscreen>
- Watch: Inference with SAHI (Slicing Aided Hyper Inference) using Ultralytics YOLOv8
+ Watch: Inference with SAHI (Slicing Aided Hyper Inference) using Ultralytics YOLO11
YOLOv8 without SAHI | -YOLOv8 with SAHI | +YOLO11 without SAHI | +YOLO11 with SAHI |
---|---|---|---|
- | + | + |
@@ -22,7 +22,7 @@ The Security Alarm System Project utilizing Ultralytics YOLOv8 integrates advanc
allowfullscreen>
- Watch: Security Alarm System Project with Ultralytics YOLOv8 [Object Detection](https://www.ultralytics.com/glossary/object-detection)
+ Watch: Security Alarm System Project with Ultralytics YOLO11 Object Detection
@@ -18,12 +18,12 @@ keywords: Ultralytics YOLOv8, speed estimation, object tracking, computer vision
allowfullscreen>
- Watch: Speed Estimation using Ultralytics YOLOv8
+ Watch: Speed Estimation using Ultralytics YOLO11
@@ -18,19 +18,19 @@ Streamlit makes it simple to build and deploy interactive web applications. Comb
allowfullscreen>
- Watch: How to Use Streamlit with Ultralytics for Real-Time [Computer Vision](https://www.ultralytics.com/glossary/computer-vision-cv) in Your Browser
+ Watch: How to Use Streamlit with Ultralytics for Real-Time Computer Vision in Your Browser
@@ -38,7 +38,7 @@ Ensure you have the following prerequisites before proceeding:
pip install tritonclient[all]
```
-## Exporting YOLOv8 to ONNX Format
+## Exporting YOLO11 to ONNX Format
Before deploying the model on Triton, it must be exported to the ONNX format. ONNX (Open Neural Network Exchange) is a format that allows models to be transferred between different deep learning frameworks. Use the `export` function from the `YOLO` class:
@@ -46,7 +46,7 @@ Before deploying the model on Triton, it must be exported to the ONNX format. ON
from ultralytics import YOLO
# Load a model
-model = YOLO("yolov8n.pt") # load an official model
+model = YOLO("yolo11n.pt") # load an official model
# Export the model
onnx_file = model.export(format="onnx", dynamic=True)
@@ -141,21 +141,21 @@ subprocess.call(f"docker kill {container_id}", shell=True)
---
-By following the above steps, you can deploy and run Ultralytics YOLOv8 models efficiently on Triton Inference Server, providing a scalable and high-performance solution for deep learning inference tasks. If you face any issues or have further queries, refer to the [official Triton documentation](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html) or reach out to the Ultralytics community for support.
+By following the above steps, you can deploy and run Ultralytics YOLO11 models efficiently on Triton Inference Server, providing a scalable and high-performance solution for deep learning inference tasks. If you face any issues or have further queries, refer to the [official Triton documentation](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html) or reach out to the Ultralytics community for support.
## FAQ
-### How do I set up Ultralytics YOLOv8 with NVIDIA Triton Inference Server?
+### How do I set up Ultralytics YOLO11 with NVIDIA Triton Inference Server?
-Setting up [Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8/) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) involves a few key steps:
+Setting up [Ultralytics YOLO11](https://docs.ultralytics.com/models/yolov8/) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) involves a few key steps:
-1. **Export YOLOv8 to ONNX format**:
+1. **Export YOLO11 to ONNX format**:
```python
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n.pt") # load an official model
+ model = YOLO("yolo11n.pt") # load an official model
# Export the model to ONNX format
onnx_file = model.export(format="onnx", dynamic=True)
@@ -209,21 +209,21 @@ Setting up [Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8/) wit
time.sleep(1)
```
-This setup can help you efficiently deploy YOLOv8 models at scale on Triton Inference Server for high-performance AI model inference.
+This setup can help you efficiently deploy YOLO11 models at scale on Triton Inference Server for high-performance AI model inference.
-### What benefits does using Ultralytics YOLOv8 with NVIDIA Triton Inference Server offer?
+### What benefits does using Ultralytics YOLO11 with NVIDIA Triton Inference Server offer?
-Integrating [Ultralytics YOLOv8](../models/yolov8.md) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) provides several advantages:
+Integrating [Ultralytics YOLO11](../models/yolov8.md) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) provides several advantages:
- **Scalable AI Inference**: Triton allows serving multiple models from a single server instance, supporting dynamic model loading and unloading, making it highly scalable for diverse AI workloads.
- **High Performance**: Optimized for NVIDIA GPUs, Triton Inference Server ensures high-speed inference operations, perfect for real-time applications such as [object detection](https://www.ultralytics.com/glossary/object-detection).
- **Ensemble and Model Versioning**: Triton's ensemble mode enables combining multiple models to improve results, and its model versioning supports A/B testing and rolling updates.
-For detailed instructions on setting up and running YOLOv8 with Triton, you can refer to the [setup guide](#setting-up-triton-model-repository).
+For detailed instructions on setting up and running YOLO11 with Triton, you can refer to the [setup guide](#setting-up-triton-model-repository).
-### Why should I export my YOLOv8 model to ONNX format before using Triton Inference Server?
+### Why should I export my YOLO11 model to ONNX format before using Triton Inference Server?
-Using ONNX (Open Neural Network Exchange) format for your [Ultralytics YOLOv8](../models/yolov8.md) model before deploying it on [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) offers several key benefits:
+Using ONNX (Open Neural Network Exchange) format for your [Ultralytics YOLO11](../models/yolov8.md) model before deploying it on [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) offers several key benefits:
- **Interoperability**: ONNX format supports transfer between different deep learning frameworks (such as PyTorch, TensorFlow), ensuring broader compatibility.
- **Optimization**: Many deployment environments, including Triton, optimize for ONNX, enabling faster inference and better performance.
@@ -234,15 +234,15 @@ To export your model, use:
```python
from ultralytics import YOLO
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
onnx_file = model.export(format="onnx", dynamic=True)
```
You can follow the steps in the [exporting guide](../modes/export.md) to complete the process.
-### Can I run inference using the Ultralytics YOLOv8 model on Triton Inference Server?
+### Can I run inference using the Ultralytics YOLO11 model on Triton Inference Server?
-Yes, you can run inference using the [Ultralytics YOLOv8](../models/yolov8.md) model on [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server). Once your model is set up in the Triton Model Repository and the server is running, you can load and run inference on your model as follows:
+Yes, you can run inference using the [Ultralytics YOLO11](../models/yolov8.md) model on [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server). Once your model is set up in the Triton Model Repository and the server is running, you can load and run inference on your model as follows:
```python
from ultralytics import YOLO
@@ -254,14 +254,14 @@ model = YOLO("http://localhost:8000/yolo", task="detect")
results = model("path/to/image.jpg")
```
-For an in-depth guide on setting up and running Triton Server with YOLOv8, refer to the [running triton inference server](#running-triton-inference-server) section.
+For an in-depth guide on setting up and running Triton Server with YOLO11, refer to the [running triton inference server](#running-triton-inference-server) section.
-### How does Ultralytics YOLOv8 compare to [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) and PyTorch models for deployment?
+### How does Ultralytics YOLO11 compare to [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) and PyTorch models for deployment?
-[Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8/) offers several unique advantages compared to TensorFlow and PyTorch models for deployment:
+[Ultralytics YOLO11](https://docs.ultralytics.com/models/yolov8/) offers several unique advantages compared to TensorFlow and PyTorch models for deployment:
-- **Real-time Performance**: Optimized for real-time object detection tasks, YOLOv8 provides state-of-the-art [accuracy](https://www.ultralytics.com/glossary/accuracy) and speed, making it ideal for applications requiring live video analytics.
-- **Ease of Use**: YOLOv8 integrates seamlessly with Triton Inference Server and supports diverse export formats (ONNX, TensorRT, CoreML), making it flexible for various deployment scenarios.
-- **Advanced Features**: YOLOv8 includes features like dynamic model loading, model versioning, and ensemble inference, which are crucial for scalable and reliable AI deployments.
+- **Real-time Performance**: Optimized for real-time object detection tasks, YOLO11 provides state-of-the-art [accuracy](https://www.ultralytics.com/glossary/accuracy) and speed, making it ideal for applications requiring live video analytics.
+- **Ease of Use**: YOLO11 integrates seamlessly with Triton Inference Server and supports diverse export formats (ONNX, TensorRT, CoreML), making it flexible for various deployment scenarios.
+- **Advanced Features**: YOLO11 includes features like dynamic model loading, model versioning, and ensemble inference, which are crucial for scalable and reliable AI deployments.
For more details, compare the deployment options in the [model deployment guide](../modes/export.md).
diff --git a/docs/en/guides/view-results-in-terminal.md b/docs/en/guides/view-results-in-terminal.md
index cb6ce1c087..2d8ff4b647 100644
--- a/docs/en/guides/view-results-in-terminal.md
+++ b/docs/en/guides/view-results-in-terminal.md
@@ -47,7 +47,7 @@ The VSCode compatible protocols for viewing images using the integrated terminal
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n.pt")
+ model = YOLO("yolo11n.pt")
# Run inference on an image
results = model.predict(source="ultralytics/assets/bus.jpg")
@@ -111,7 +111,7 @@ from sixel import SixelWriter
from ultralytics import YOLO
# Load a model
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
# Run inference on an image
results = model.predict(source="ultralytics/assets/bus.jpg")
@@ -164,7 +164,7 @@ To view YOLO inference results in a VSCode terminal on macOS or Linux, follow th
```python
from ultralytics import YOLO
- model = YOLO("yolov8n.pt")
+ model = YOLO("yolo11n.pt")
results = model.predict(source="path_to_image")
plot = results[0].plot()
```
diff --git a/docs/en/guides/vision-eye.md b/docs/en/guides/vision-eye.md
index 48cff2746d..f374388c9c 100644
--- a/docs/en/guides/vision-eye.md
+++ b/docs/en/guides/vision-eye.md
@@ -1,23 +1,23 @@
---
comments: true
-description: Discover VisionEye's object mapping and tracking powered by Ultralytics YOLOv8. Simulate human eye precision, track objects, and calculate distances effortlessly.
-keywords: VisionEye, YOLOv8, Ultralytics, object mapping, object tracking, distance calculation, computer vision, AI, machine learning, Python, tutorial
+description: Discover VisionEye's object mapping and tracking powered by Ultralytics YOLO11. Simulate human eye precision, track objects, and calculate distances effortlessly.
+keywords: VisionEye, YOLO11, Ultralytics, object mapping, object tracking, distance calculation, computer vision, AI, machine learning, Python, tutorial
---
-# VisionEye View Object Mapping using Ultralytics YOLOv8 🚀
+# VisionEye View Object Mapping using Ultralytics YOLO11 🚀
## What is VisionEye Object Mapping?
-[Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics/) VisionEye offers the capability for computers to identify and pinpoint objects, simulating the observational [precision](https://www.ultralytics.com/glossary/precision) of the human eye. This functionality enables computers to discern and focus on specific objects, much like the way the human eye observes details from a particular viewpoint.
+[Ultralytics YOLO11](https://github.com/ultralytics/ultralytics/) VisionEye offers the capability for computers to identify and pinpoint objects, simulating the observational [precision](https://www.ultralytics.com/glossary/precision) of the human eye. This functionality enables computers to discern and focus on specific objects, much like the way the human eye observes details from a particular viewpoint.
## Samples
| VisionEye View | VisionEye View With Object Tracking | VisionEye View With Distance Calculation |
| :----------------------------------------------------------------------------------------------------------------------------------------------------------: | :---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------: | :----------------------------------------------------------------------------------------------------------------------------------------------------------------------: |
-| ![VisionEye View Object Mapping using Ultralytics YOLOv8](https://github.com/ultralytics/docs/releases/download/0/visioneye-view-object-mapping-yolov8.avif) | ![VisionEye View Object Mapping with Object Tracking using Ultralytics YOLOv8](https://github.com/ultralytics/docs/releases/download/0/visioneye-object-mapping-with-tracking.avif) | ![VisionEye View with Distance Calculation using Ultralytics YOLOv8](https://github.com/ultralytics/docs/releases/download/0/visioneye-distance-calculation-yolov8.avif) |
-| VisionEye View Object Mapping using Ultralytics YOLOv8 | VisionEye View Object Mapping with Object Tracking using Ultralytics YOLOv8 | VisionEye View with Distance Calculation using Ultralytics YOLOv8 |
+| ![VisionEye View Object Mapping using Ultralytics YOLO11](https://github.com/ultralytics/docs/releases/download/0/visioneye-view-object-mapping-yolov8.avif) | ![VisionEye View Object Mapping with Object Tracking using Ultralytics YOLO11](https://github.com/ultralytics/docs/releases/download/0/visioneye-object-mapping-with-tracking.avif) | ![VisionEye View with Distance Calculation using Ultralytics YOLO11](https://github.com/ultralytics/docs/releases/download/0/visioneye-distance-calculation-yolov8.avif) |
+| VisionEye View Object Mapping using Ultralytics YOLO11 | VisionEye View Object Mapping with Object Tracking using Ultralytics YOLO11 | VisionEye View with Distance Calculation using Ultralytics YOLO11 |
-!!! example "VisionEye Object Mapping using YOLOv8"
+!!! example "VisionEye Object Mapping using YOLO11"
=== "VisionEye Object Mapping"
@@ -27,7 +27,7 @@ keywords: VisionEye, YOLOv8, Ultralytics, object mapping, object tracking, dista
from ultralytics import YOLO
from ultralytics.utils.plotting import Annotator, colors
- model = YOLO("yolov8n.pt")
+ model = YOLO("yolo11n.pt")
names = model.model.names
cap = cv2.VideoCapture("path/to/video/file.mp4")
w, h, fps = (int(cap.get(x)) for x in (cv2.CAP_PROP_FRAME_WIDTH, cv2.CAP_PROP_FRAME_HEIGHT, cv2.CAP_PROP_FPS))
@@ -71,7 +71,7 @@ keywords: VisionEye, YOLOv8, Ultralytics, object mapping, object tracking, dista
from ultralytics import YOLO
from ultralytics.utils.plotting import Annotator, colors
- model = YOLO("yolov8n.pt")
+ model = YOLO("yolo11n.pt")
cap = cv2.VideoCapture("path/to/video/file.mp4")
w, h, fps = (int(cap.get(x)) for x in (cv2.CAP_PROP_FRAME_WIDTH, cv2.CAP_PROP_FRAME_HEIGHT, cv2.CAP_PROP_FPS))
@@ -118,7 +118,7 @@ keywords: VisionEye, YOLOv8, Ultralytics, object mapping, object tracking, dista
from ultralytics import YOLO
from ultralytics.utils.plotting import Annotator
- model = YOLO("yolov8s.pt")
+ model = YOLO("yolo11n.pt")
cap = cv2.VideoCapture("Path/to/video/file.mp4")
w, h, fps = (int(cap.get(x)) for x in (cv2.CAP_PROP_FRAME_WIDTH, cv2.CAP_PROP_FRAME_HEIGHT, cv2.CAP_PROP_FPS))
@@ -180,16 +180,16 @@ For any inquiries, feel free to post your questions in the [Ultralytics Issue Se
## FAQ
-### How do I start using VisionEye Object Mapping with Ultralytics YOLOv8?
+### How do I start using VisionEye Object Mapping with Ultralytics YOLO11?
-To start using VisionEye Object Mapping with Ultralytics YOLOv8, first, you'll need to install the Ultralytics YOLO package via pip. Then, you can use the sample code provided in the documentation to set up [object detection](https://www.ultralytics.com/glossary/object-detection) with VisionEye. Here's a simple example to get you started:
+To start using VisionEye Object Mapping with Ultralytics YOLO11, first, you'll need to install the Ultralytics YOLO package via pip. Then, you can use the sample code provided in the documentation to set up [object detection](https://www.ultralytics.com/glossary/object-detection) with VisionEye. Here's a simple example to get you started:
```python
import cv2
from ultralytics import YOLO
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
cap = cv2.VideoCapture("path/to/video/file.mp4")
while True:
@@ -210,12 +210,12 @@ cap.release()
cv2.destroyAllWindows()
```
-### What are the key features of VisionEye's object tracking capability using Ultralytics YOLOv8?
+### What are the key features of VisionEye's object tracking capability using Ultralytics YOLO11?
-VisionEye's object tracking with Ultralytics YOLOv8 allows users to follow the movement of objects within a video frame. Key features include:
+VisionEye's object tracking with Ultralytics YOLO11 allows users to follow the movement of objects within a video frame. Key features include:
1. **Real-Time Object Tracking**: Keeps up with objects as they move.
-2. **Object Identification**: Utilizes YOLOv8's powerful detection algorithms.
+2. **Object Identification**: Utilizes YOLO11's powerful detection algorithms.
3. **Distance Calculation**: Calculates distances between objects and specified points.
4. **Annotation and Visualization**: Provides visual markers for tracked objects.
@@ -226,7 +226,7 @@ import cv2
from ultralytics import YOLO
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
cap = cv2.VideoCapture("path/to/video/file.mp4")
while True:
@@ -249,9 +249,9 @@ cv2.destroyAllWindows()
For a comprehensive guide, visit the [VisionEye Object Mapping with Object Tracking](#samples).
-### How can I calculate distances with VisionEye's YOLOv8 model?
+### How can I calculate distances with VisionEye's YOLO11 model?
-Distance calculation with VisionEye and Ultralytics YOLOv8 involves determining the distance of detected objects from a specified point in the frame. It enhances spatial analysis capabilities, useful in applications such as autonomous driving and surveillance.
+Distance calculation with VisionEye and Ultralytics YOLO11 involves determining the distance of detected objects from a specified point in the frame. It enhances spatial analysis capabilities, useful in applications such as autonomous driving and surveillance.
Here's a simplified example:
@@ -262,7 +262,7 @@ import cv2
from ultralytics import YOLO
-model = YOLO("yolov8s.pt")
+model = YOLO("yolo11n.pt")
cap = cv2.VideoCapture("path/to/video/file.mp4")
center_point = (0, 480) # Example center point
pixel_per_meter = 10
@@ -290,19 +290,19 @@ cv2.destroyAllWindows()
For detailed instructions, refer to the [VisionEye with Distance Calculation](#samples).
-### Why should I use Ultralytics YOLOv8 for object mapping and tracking?
+### Why should I use Ultralytics YOLO11 for object mapping and tracking?
-Ultralytics YOLOv8 is renowned for its speed, [accuracy](https://www.ultralytics.com/glossary/accuracy), and ease of integration, making it a top choice for object mapping and tracking. Key advantages include:
+Ultralytics YOLO11 is renowned for its speed, [accuracy](https://www.ultralytics.com/glossary/accuracy), and ease of integration, making it a top choice for object mapping and tracking. Key advantages include:
1. **State-of-the-art Performance**: Delivers high accuracy in real-time object detection.
2. **Flexibility**: Supports various tasks such as detection, tracking, and distance calculation.
3. **Community and Support**: Extensive documentation and active GitHub community for troubleshooting and enhancements.
4. **Ease of Use**: Intuitive API simplifies complex tasks, allowing for rapid deployment and iteration.
-For more information on applications and benefits, check out the [Ultralytics YOLOv8 documentation](https://docs.ultralytics.com/models/yolov8/).
+For more information on applications and benefits, check out the [Ultralytics YOLO11 documentation](https://docs.ultralytics.com/models/yolov8/).
### How can I integrate VisionEye with other [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) tools like Comet or ClearML?
-Ultralytics YOLOv8 can integrate seamlessly with various machine learning tools like Comet and ClearML, enhancing experiment tracking, collaboration, and reproducibility. Follow the detailed guides on [how to use YOLOv5 with Comet](https://www.ultralytics.com/blog/how-to-use-yolov5-with-comet) and [integrate YOLOv8 with ClearML](https://docs.ultralytics.com/integrations/clearml/) to get started.
+Ultralytics YOLO11 can integrate seamlessly with various machine learning tools like Comet and ClearML, enhancing experiment tracking, collaboration, and reproducibility. Follow the detailed guides on [how to use YOLOv5 with Comet](https://www.ultralytics.com/blog/how-to-use-yolov5-with-comet) and [integrate YOLO11 with ClearML](https://docs.ultralytics.com/integrations/clearml/) to get started.
For further exploration and integration examples, check our [Ultralytics Integrations Guide](https://docs.ultralytics.com/integrations/).
diff --git a/docs/en/guides/workouts-monitoring.md b/docs/en/guides/workouts-monitoring.md
index 4585631683..af996894b3 100644
--- a/docs/en/guides/workouts-monitoring.md
+++ b/docs/en/guides/workouts-monitoring.md
@@ -1,12 +1,12 @@
---
comments: true
-description: Optimize your fitness routine with real-time workouts monitoring using Ultralytics YOLOv8. Track and improve your exercise form and performance.
-keywords: workouts monitoring, Ultralytics YOLOv8, pose estimation, fitness tracking, exercise assessment, real-time feedback, exercise form, performance metrics
+description: Optimize your fitness routine with real-time workouts monitoring using Ultralytics YOLO11. Track and improve your exercise form and performance.
+keywords: workouts monitoring, Ultralytics YOLO11, pose estimation, fitness tracking, exercise assessment, real-time feedback, exercise form, performance metrics
---
-# Workouts Monitoring using Ultralytics YOLOv8
+# Workouts Monitoring using Ultralytics YOLO11
-Monitoring workouts through pose estimation with [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics/) enhances exercise assessment by accurately tracking key body landmarks and joints in real-time. This technology provides instant feedback on exercise form, tracks workout routines, and measures performance metrics, optimizing training sessions for users and trainers alike.
+Monitoring workouts through pose estimation with [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics/) enhances exercise assessment by accurately tracking key body landmarks and joints in real-time. This technology provides instant feedback on exercise form, tracks workout routines, and measures performance metrics, optimizing training sessions for users and trainers alike.
@@ -16,7 +16,7 @@ Monitoring workouts through pose estimation with [Ultralytics YOLOv8](https://gi
allowfullscreen>
- Watch: Workouts Monitoring using Ultralytics YOLOv8 | Pushups, Pullups, Ab Workouts
+ Watch: Workouts Monitoring using Ultralytics YOLO11 | Pushups, Pullups, Ab Workouts
@@ -22,7 +22,7 @@ This guide serves as a comprehensive aid for troubleshooting common issues encou
allowfullscreen>
- Watch: Ultralytics YOLOv8 Common Issues | Installation Errors, Model Training Issues
+ Watch: Ultralytics YOLO11 Common Issues | Installation Errors, Model Training Issues
@@ -18,12 +18,12 @@ Performance metrics are key tools to evaluate the [accuracy](https://www.ultraly
allowfullscreen>
- Watch: Ultralytics YOLOv8 Performance Metrics | MAP, F1 Score, [Precision](https://www.ultralytics.com/glossary/precision), IoU & Accuracy
+ Watch: Ultralytics YOLO11 Performance Metrics | MAP, F1 Score, Precision, IoU & Accuracy
@@ -81,10 +81,10 @@ Explore the Ultralytics Docs, a comprehensive resource designed to help you unde
- [YOLOv5](https://github.com/ultralytics/yolov5) further improved the model's performance and added new features such as hyperparameter optimization, integrated experiment tracking and automatic export to popular export formats.
- [YOLOv6](https://github.com/meituan/YOLOv6) was open-sourced by [Meituan](https://about.meituan.com/) in 2022 and is in use in many of the company's autonomous delivery robots.
- [YOLOv7](https://github.com/WongKinYiu/yolov7) added additional tasks such as pose estimation on the COCO keypoints dataset.
-- [YOLOv8](https://github.com/ultralytics/ultralytics) is the latest version of YOLO by Ultralytics. As a cutting-edge, state-of-the-art (SOTA) model, YOLOv8 builds on the success of previous versions, introducing new features and improvements for enhanced performance, flexibility, and efficiency. YOLOv8 supports a full range of vision AI tasks, including [detection](tasks/detect.md), [segmentation](tasks/segment.md), [pose estimation](tasks/pose.md), [tracking](modes/track.md), and [classification](tasks/classify.md). This versatility allows users to leverage YOLOv8's capabilities across diverse applications and domains.
+- [YOLOv8](https://github.com/ultralytics/ultralytics) released in 2023 by Ultralytics. YOLOv8 introduced new features and improvements for enhanced performance, flexibility, and efficiency, supporting a full range of vision AI tasks,
- [YOLOv9](models/yolov9.md) introduces innovative methods like Programmable Gradient Information (PGI) and the Generalized Efficient Layer Aggregation Network (GELAN).
- [YOLOv10](models/yolov10.md) is created by researchers from [Tsinghua University](https://www.tsinghua.edu.cn/en/) using the [Ultralytics](https://www.ultralytics.com/) [Python package](https://pypi.org/project/ultralytics/). This version provides real-time [object detection](tasks/detect.md) advancements by introducing an End-to-End head that eliminates Non-Maximum Suppression (NMS) requirements.
-- **[YOLO11](models/yolo11.md) NEW 🚀**: Ultralytics' latest YOLO models delivering state-of-the-art (SOTA) performance across multiple tasks.
+- **[YOLO11](models/yolo11.md) NEW 🚀**: Ultralytics' latest YOLO models delivering state-of-the-art (SOTA) performance across multiple tasks, including [detection](tasks/detect.md), [segmentation](tasks/segment.md), [pose estimation](tasks/pose.md), [tracking](modes/track.md), and [classification](tasks/classify.md), leverage capabilities across diverse AI applications and domains.
## YOLO Licenses: How is Ultralytics YOLO licensed?
@@ -103,11 +103,15 @@ Ultralytics YOLO is the latest advancement in the acclaimed YOLO (You Only Look
### How can I get started with YOLO installation and setup?
-Getting started with YOLO is quick and straightforward. You can install the Ultralytics package using pip and get up and running in minutes. Here's a basic installation command:
+Getting started with YOLO is quick and straightforward. You can install the Ultralytics package using [pip](https://pypi.org/project/ultralytics/) and get up and running in minutes. Here's a basic installation command:
-```bash
-pip install ultralytics
-```
+!!! example
+
+ === "CLI"
+
+ ```bash
+ pip install ultralytics
+ ```
For a comprehensive step-by-step guide, visit our [quickstart guide](quickstart.md). This resource will help you with installation instructions, initial setup, and running your first model.
@@ -119,11 +123,28 @@ Training a custom YOLO model on your dataset involves a few detailed steps:
2. Configure the training parameters in a YAML file.
3. Use the `yolo train` command to start training.
-Here's an example command:
+Here's example code:
+
+!!! example
+
+ === "Python"
+
+ ```python
+ from ultralytics import YOLO
+
+ # Load a pre-trained YOLO model (you can choose n, s, m, l, or x versions)
+ model = YOLO("yolo11n.pt")
+
+ # Start training on your custom dataset
+ model.train(data="path/to/dataset.yaml", epochs=100, imgsz=640)
+ ```
-```bash
-yolo train model=yolo11n.pt data=coco128.yaml epochs=100 imgsz=640
-```
+ === "CLI"
+
+ ```bash
+ # Train a YOLO model from the command line
+ yolo train data=path/to/dataset.yaml epochs=100 imgsz=640
+ ```
For a detailed walkthrough, check out our [Train a Model](modes/train.md) guide, which includes examples and tips for optimizing your training process.
@@ -140,8 +161,27 @@ For more details, visit our [Licensing](https://www.ultralytics.com/license) pag
Ultralytics YOLO supports efficient and customizable multi-object tracking. To utilize tracking capabilities, you can use the `yolo track` command as shown below:
-```bash
-yolo track model=yolo11n.pt source=video.mp4
-```
+!!! example
+
+ === "Python"
+
+ ```python
+ from ultralytics import YOLO
+
+ # Load a pre-trained YOLO model
+ model = YOLO("yolo11n.pt")
+
+ # Start tracking objects in a video
+ # You can also use live video streams or webcam input
+ model.track(source="path/to/video.mp4")
+ ```
+
+ === "CLI"
+
+ ```bash
+ # Perform object tracking on a video from the command line
+ # You can specify different sources like webcam (0) or RTSP streams
+ yolo track source=path/to/video.mp4
+ ```
For a detailed guide on setting up and running object tracking, check our [tracking mode](modes/track.md) documentation, which explains the configuration and practical applications in real-time scenarios.
diff --git a/docs/en/integrations/amazon-sagemaker.md b/docs/en/integrations/amazon-sagemaker.md
index 9a82037ac2..366a4f1388 100644
--- a/docs/en/integrations/amazon-sagemaker.md
+++ b/docs/en/integrations/amazon-sagemaker.md
@@ -1,14 +1,14 @@
---
comments: true
-description: Learn step-by-step how to deploy Ultralytics' YOLOv8 on Amazon SageMaker Endpoints, from setup to testing, for powerful real-time inference with AWS services.
-keywords: YOLOv8, Amazon SageMaker, AWS, Ultralytics, machine learning, computer vision, model deployment, AWS CloudFormation, AWS CDK, real-time inference
+description: Learn step-by-step how to deploy Ultralytics' YOLO11 on Amazon SageMaker Endpoints, from setup to testing, for powerful real-time inference with AWS services.
+keywords: YOLO11, Amazon SageMaker, AWS, Ultralytics, machine learning, computer vision, model deployment, AWS CloudFormation, AWS CDK, real-time inference
---
-# A Guide to Deploying YOLOv8 on Amazon SageMaker Endpoints
+# A Guide to Deploying YOLO11 on Amazon SageMaker Endpoints
-Deploying advanced [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models like [Ultralytics' YOLOv8](https://github.com/ultralytics/ultralytics) on Amazon SageMaker Endpoints opens up a wide range of possibilities for various [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) applications. The key to effectively using these models lies in understanding their setup, configuration, and deployment processes. YOLOv8 becomes even more powerful when integrated seamlessly with Amazon SageMaker, a robust and scalable machine learning service by AWS.
+Deploying advanced [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models like [Ultralytics' YOLO11](https://github.com/ultralytics/ultralytics) on Amazon SageMaker Endpoints opens up a wide range of possibilities for various [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) applications. The key to effectively using these models lies in understanding their setup, configuration, and deployment processes. YOLO11 becomes even more powerful when integrated seamlessly with Amazon SageMaker, a robust and scalable machine learning service by AWS.
-This guide will take you through the process of deploying YOLOv8 [PyTorch](https://www.ultralytics.com/glossary/pytorch) models on Amazon SageMaker Endpoints step by step. You'll learn the essentials of preparing your AWS environment, configuring the model appropriately, and using tools like AWS CloudFormation and the AWS Cloud Development Kit (CDK) for deployment.
+This guide will take you through the process of deploying YOLO11 [PyTorch](https://www.ultralytics.com/glossary/pytorch) models on Amazon SageMaker Endpoints step by step. You'll learn the essentials of preparing your AWS environment, configuring the model appropriately, and using tools like AWS CloudFormation and the AWS Cloud Development Kit (CDK) for deployment.
## Amazon SageMaker
@@ -18,9 +18,9 @@ This guide will take you through the process of deploying YOLOv8 [PyTorch](https
[Amazon SageMaker](https://aws.amazon.com/sagemaker/) is a machine learning service from Amazon Web Services (AWS) that simplifies the process of building, training, and deploying machine learning models. It provides a broad range of tools for handling various aspects of machine learning workflows. This includes automated features for tuning models, options for training models at scale, and straightforward methods for deploying models into production. SageMaker supports popular machine learning frameworks, offering the flexibility needed for diverse projects. Its features also cover data labeling, workflow management, and performance analysis.
-## Deploying YOLOv8 on Amazon SageMaker Endpoints
+## Deploying YOLO11 on Amazon SageMaker Endpoints
-Deploying YOLOv8 on Amazon SageMaker lets you use its managed environment for real-time inference and take advantage of features like autoscaling. Take a look at the AWS architecture below.
+Deploying YOLO11 on Amazon SageMaker lets you use its managed environment for real-time inference and take advantage of features like autoscaling. Take a look at the AWS architecture below.
@@ -40,9 +40,9 @@ First, ensure you have the following prerequisites in place: - Adequate Service Quota: Confirm that you have sufficient quotas for two separate resources in Amazon SageMaker: one for `ml.m5.4xlarge` for endpoint usage and another for `ml.m5.4xlarge` for notebook instance usage. Each of these requires a minimum of one quota value. If your current quotas are below this requirement, it's important to request an increase for each. You can request a quota increase by following the detailed instructions in the [AWS Service Quotas documentation](https://docs.aws.amazon.com/servicequotas/latest/userguide/request-quota-increase.html#quota-console-increase). -### Step 2: Clone the YOLOv8 SageMaker Repository +### Step 2: Clone the YOLO11 SageMaker Repository -The next step is to clone the specific AWS repository that contains the resources for deploying YOLOv8 on SageMaker. This repository, hosted on GitHub, includes the necessary CDK scripts and configuration files. +The next step is to clone the specific AWS repository that contains the resources for deploying YOLO11 on SageMaker. This repository, hosted on GitHub, includes the necessary CDK scripts and configuration files. - Clone the GitHub Repository: Execute the following command in your terminal to clone the host-yolov8-on-sagemaker-endpoint repository: @@ -104,11 +104,11 @@ cdk bootstrap cdk deploy ``` -### Step 5: Deploy the YOLOv8 Model +### Step 5: Deploy the YOLO Model -Before diving into the deployment instructions, be sure to check out the range of [YOLOv8 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. +Before diving into the deployment instructions, be sure to check out the range of [YOLO11 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. -After creating the AWS CloudFormation Stack, the next step is to deploy YOLOv8. +After creating the AWS CloudFormation Stack, the next step is to deploy YOLO11. - Open the Notebook Instance: Go to the AWS Console and navigate to the Amazon SageMaker service. Select "Notebook Instances" from the dashboard, then locate the notebook instance that was created by your CDK deployment script. Open the notebook instance to access the Jupyter environment. @@ -136,18 +136,18 @@ def output_fn(prediction_output): return json.dumps(infer) ``` -- Deploy the Endpoint Using 1_DeployEndpoint.ipynb: In the Jupyter environment, open the 1_DeployEndpoint.ipynb notebook located in the sm-notebook directory. Follow the instructions in the notebook and run the cells to download the YOLOv8 model, package it with the updated inference code, and upload it to an Amazon S3 bucket. The notebook will guide you through creating and deploying a SageMaker endpoint for the YOLOv8 model. +- Deploy the Endpoint Using 1_DeployEndpoint.ipynb: In the Jupyter environment, open the 1_DeployEndpoint.ipynb notebook located in the sm-notebook directory. Follow the instructions in the notebook and run the cells to download the YOLO11 model, package it with the updated inference code, and upload it to an Amazon S3 bucket. The notebook will guide you through creating and deploying a SageMaker endpoint for the YOLO11 model. ### Step 6: Testing Your Deployment -Now that your YOLOv8 model is deployed, it's important to test its performance and functionality. +Now that your YOLO11 model is deployed, it's important to test its performance and functionality. - Open the Test Notebook: In the same Jupyter environment, locate and open the 2_TestEndpoint.ipynb notebook, also in the sm-notebook directory. - Run the Test Notebook: Follow the instructions within the notebook to test the deployed SageMaker endpoint. This includes sending an image to the endpoint and running inferences. Then, you'll plot the output to visualize the model's performance and [accuracy](https://www.ultralytics.com/glossary/accuracy), as shown below.
- +
- Clean-Up Resources: The test notebook will also guide you through the process of cleaning up the endpoint and the hosted model. This is an important step to manage costs and resources effectively, especially if you do not plan to use the deployed model immediately. @@ -160,24 +160,24 @@ After testing, continuous monitoring and management of your deployed model are e - Manage the Endpoint: Use the SageMaker console for ongoing management of the endpoint. This includes scaling, updating, or redeploying the model as required. -By completing these steps, you will have successfully deployed and tested a YOLOv8 model on Amazon SageMaker Endpoints. This process not only equips you with practical experience in using AWS services for machine learning deployment but also lays the foundation for deploying other advanced models in the future. +By completing these steps, you will have successfully deployed and tested a YOLO11 model on Amazon SageMaker Endpoints. This process not only equips you with practical experience in using AWS services for machine learning deployment but also lays the foundation for deploying other advanced models in the future. ## Summary -This guide took you step by step through deploying YOLOv8 on Amazon SageMaker Endpoints using AWS CloudFormation and the AWS Cloud Development Kit (CDK). The process includes cloning the necessary GitHub repository, setting up the CDK environment, deploying the model using AWS services, and testing its performance on SageMaker. +This guide took you step by step through deploying YOLO11 on Amazon SageMaker Endpoints using AWS CloudFormation and the AWS Cloud Development Kit (CDK). The process includes cloning the necessary GitHub repository, setting up the CDK environment, deploying the model using AWS services, and testing its performance on SageMaker. For more technical details, refer to [this article](https://aws.amazon.com/blogs/machine-learning/hosting-yolov8-pytorch-model-on-amazon-sagemaker-endpoints/) on the AWS Machine Learning Blog. You can also check out the official [Amazon SageMaker Documentation](https://docs.aws.amazon.com/sagemaker/latest/dg/realtime-endpoints.html) for more insights into various features and functionalities. -Are you interested in learning more about different YOLOv8 integrations? Visit the [Ultralytics integrations guide page](../integrations/index.md) to discover additional tools and capabilities that can enhance your machine-learning projects. +Are you interested in learning more about different YOLO11 integrations? Visit the [Ultralytics integrations guide page](../integrations/index.md) to discover additional tools and capabilities that can enhance your machine-learning projects. ## FAQ -### How do I deploy the Ultralytics YOLOv8 model on Amazon SageMaker Endpoints? +### How do I deploy the Ultralytics YOLO11 model on Amazon SageMaker Endpoints? -To deploy the Ultralytics YOLOv8 model on Amazon SageMaker Endpoints, follow these steps: +To deploy the Ultralytics YOLO11 model on Amazon SageMaker Endpoints, follow these steps: 1. **Set Up Your AWS Environment**: Ensure you have an AWS Account, IAM roles with necessary permissions, and the AWS CLI configured. Install AWS CDK if not already done (refer to the [AWS CDK instructions](https://docs.aws.amazon.com/cdk/v2/guide/getting_started.html#getting_started_install)). -2. **Clone the YOLOv8 SageMaker Repository**: +2. **Clone the YOLO11 SageMaker Repository**: ```bash git clone https://github.com/aws-samples/host-yolov8-on-sagemaker-endpoint.git cd host-yolov8-on-sagemaker-endpoint/yolov8-pytorch-cdk @@ -196,11 +196,11 @@ To deploy the Ultralytics YOLOv8 model on Amazon SageMaker Endpoints, follow the cdk deploy ``` -For further details, review the [documentation section](#step-5-deploy-the-yolov8-model). +For further details, review the [documentation section](#step-5-deploy-the-yolo-model). -### What are the prerequisites for deploying YOLOv8 on Amazon SageMaker? +### What are the prerequisites for deploying YOLO11 on Amazon SageMaker? -To deploy YOLOv8 on Amazon SageMaker, ensure you have the following prerequisites: +To deploy YOLO11 on Amazon SageMaker, ensure you have the following prerequisites: 1. **AWS Account**: Active AWS account ([sign up here](https://aws.amazon.com/)). 2. **IAM Roles**: Configured IAM roles with permissions for SageMaker, CloudFormation, and Amazon S3. @@ -210,9 +210,9 @@ To deploy YOLOv8 on Amazon SageMaker, ensure you have the following prerequisite For detailed setup, refer to [this section](#step-1-setup-your-aws-environment). -### Why should I use Ultralytics YOLOv8 on Amazon SageMaker? +### Why should I use Ultralytics YOLO11 on Amazon SageMaker? -Using Ultralytics YOLOv8 on Amazon SageMaker offers several advantages: +Using Ultralytics YOLO11 on Amazon SageMaker offers several advantages: 1. **Scalability and Management**: SageMaker provides a managed environment with features like autoscaling, which helps in real-time inference needs. 2. **Integration with AWS Services**: Seamlessly integrate with other AWS services, such as S3 for data storage, CloudFormation for infrastructure as code, and CloudWatch for monitoring. @@ -221,9 +221,9 @@ Using Ultralytics YOLOv8 on Amazon SageMaker offers several advantages: Explore more about the advantages of using SageMaker in the [introduction section](#amazon-sagemaker). -### Can I customize the inference logic for YOLOv8 on Amazon SageMaker? +### Can I customize the inference logic for YOLO11 on Amazon SageMaker? -Yes, you can customize the inference logic for YOLOv8 on Amazon SageMaker: +Yes, you can customize the inference logic for YOLO11 on Amazon SageMaker: 1. **Modify `inference.py`**: Locate and customize the `output_fn` function in the `inference.py` file to tailor output formats. @@ -243,11 +243,11 @@ Yes, you can customize the inference logic for YOLOv8 on Amazon SageMaker: 2. **Deploy Updated Model**: Ensure you redeploy the model using Jupyter notebooks provided (`1_DeployEndpoint.ipynb`) to include these changes. -Refer to the [detailed steps](#step-5-deploy-the-yolov8-model) for deploying the modified model. +Refer to the [detailed steps](#step-5-deploy-the-yolo-model) for deploying the modified model. -### How can I test the deployed YOLOv8 model on Amazon SageMaker? +### How can I test the deployed YOLO11 model on Amazon SageMaker? -To test the deployed YOLOv8 model on Amazon SageMaker: +To test the deployed YOLO11 model on Amazon SageMaker: 1. **Open the Test Notebook**: Locate the `2_TestEndpoint.ipynb` notebook in the SageMaker Jupyter environment. 2. **Run the Notebook**: Follow the notebook's instructions to send an image to the endpoint, perform inference, and display results. diff --git a/docs/en/integrations/clearml.md b/docs/en/integrations/clearml.md index 5245d88740..465f2fa8d1 100644 --- a/docs/en/integrations/clearml.md +++ b/docs/en/integrations/clearml.md @@ -1,14 +1,14 @@ --- comments: true -description: Discover how to integrate YOLOv8 with ClearML to streamline your MLOps workflow, automate experiments, and enhance model management effortlessly. -keywords: YOLOv8, ClearML, MLOps, Ultralytics, machine learning, object detection, model training, automation, experiment management +description: Discover how to integrate YOLO11 with ClearML to streamline your MLOps workflow, automate experiments, and enhance model management effortlessly. +keywords: YOLO11, ClearML, MLOps, Ultralytics, machine learning, object detection, model training, automation, experiment management --- -# Training YOLOv8 with ClearML: Streamlining Your MLOps Workflow +# Training YOLO11 with ClearML: Streamlining Your MLOps Workflow MLOps bridges the gap between creating and deploying [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) models in real-world settings. It focuses on efficient deployment, scalability, and ongoing management to ensure models perform well in practical applications. -[Ultralytics YOLOv8](https://www.ultralytics.com/) effortlessly integrates with ClearML, streamlining and enhancing your [object detection](https://www.ultralytics.com/glossary/object-detection) model's training and management. This guide will walk you through the integration process, detailing how to set up ClearML, manage experiments, automate model management, and collaborate effectively. +[Ultralytics YOLO11](https://www.ultralytics.com/) effortlessly integrates with ClearML, streamlining and enhancing your [object detection](https://www.ultralytics.com/glossary/object-detection) model's training and management. This guide will walk you through the integration process, detailing how to set up ClearML, manage experiments, automate model management, and collaborate effectively. ## ClearML @@ -18,9 +18,9 @@ MLOps bridges the gap between creating and deploying [machine learning](https:// [ClearML](https://clear.ml/) is an innovative open-source MLOps platform that is skillfully designed to automate, monitor, and orchestrate machine learning workflows. Its key features include automated logging of all training and inference data for full experiment reproducibility, an intuitive web UI for easy [data visualization](https://www.ultralytics.com/glossary/data-visualization) and analysis, advanced hyperparameter [optimization algorithms](https://www.ultralytics.com/glossary/optimization-algorithm), and robust model management for efficient deployment across various platforms. -## YOLOv8 Training with ClearML +## YOLO11 Training with ClearML -You can bring automation and efficiency to your machine learning workflow by improving your training process by integrating YOLOv8 with ClearML. +You can bring automation and efficiency to your machine learning workflow by improving your training process by integrating YOLO11 with ClearML. ## Installation @@ -31,11 +31,11 @@ To install the required packages, run: === "CLI" ```bash - # Install the required packages for YOLOv8 and ClearML + # Install the required packages for YOLO11 and ClearML pip install ultralytics clearml ``` -For detailed instructions and best practices related to the installation process, be sure to check our [YOLOv8 Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, be sure to check our [YOLO11 Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ## Configuring ClearML @@ -56,7 +56,7 @@ After executing this command, visit the [ClearML Settings page](https://app.clea ## Usage -Before diving into the usage instructions, be sure to check out the range of [YOLOv8 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. +Before diving into the usage instructions, be sure to check out the range of [YOLO11 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. !!! example "Usage" @@ -70,11 +70,11 @@ Before diving into the usage instructions, be sure to check out the range of [YO # Step 1: Creating a ClearML Task task = Task.init(project_name="my_project", task_name="my_yolov8_task") - # Step 2: Selecting the YOLOv8 Model - model_variant = "yolov8n" + # Step 2: Selecting the YOLO11 Model + model_variant = "yolo11n" task.set_parameter("model_variant", model_variant) - # Step 3: Loading the YOLOv8 Model + # Step 3: Loading the YOLO11 Model model = YOLO(f"{model_variant}.pt") # Step 4: Setting Up Training Arguments @@ -91,11 +91,11 @@ Let's understand the steps showcased in the usage code snippet above. **Step 1: Creating a ClearML Task**: A new task is initialized in ClearML, specifying your project and task names. This task will track and manage your model's training. -**Step 2: Selecting the YOLOv8 Model**: The `model_variant` variable is set to 'yolov8n', one of the YOLOv8 models. This variant is then logged in ClearML for tracking. +**Step 2: Selecting the YOLO11 Model**: The `model_variant` variable is set to 'yolo11n', one of the YOLO11 models. This variant is then logged in ClearML for tracking. -**Step 3: Loading the YOLOv8 Model**: The selected YOLOv8 model is loaded using Ultralytics' YOLO class, preparing it for training. +**Step 3: Loading the YOLO11 Model**: The selected YOLO11 model is loaded using Ultralytics' YOLO class, preparing it for training. -**Step 4: Setting Up Training Arguments**: Key training arguments like the dataset (`coco8.yaml`) and the number of [epochs](https://www.ultralytics.com/glossary/epoch) (`16`) are organized in a dictionary and connected to the ClearML task. This allows for tracking and potential modification via the ClearML UI. For a detailed understanding of the model training process and best practices, refer to our [YOLOv8 Model Training guide](../modes/train.md). +**Step 4: Setting Up Training Arguments**: Key training arguments like the dataset (`coco8.yaml`) and the number of [epochs](https://www.ultralytics.com/glossary/epoch) (`16`) are organized in a dictionary and connected to the ClearML task. This allows for tracking and potential modification via the ClearML UI. For a detailed understanding of the model training process and best practices, refer to our [YOLO11 Model Training guide](../modes/train.md). **Step 5: Initiating Model Training**: The model training is started with the specified arguments. The results of the training process are captured in the `results` variable. @@ -106,7 +106,7 @@ Upon running the usage code snippet above, you can expect the following output: - A confirmation message indicating the creation of a new ClearML task, along with its unique ID. - An informational message about the script code being stored, indicating that the code execution is being tracked by ClearML. - A URL link to the ClearML results page where you can monitor the training progress and view detailed logs. -- Download progress for the YOLOv8 model and the specified dataset, followed by a summary of the model architecture and training configuration. +- Download progress for the YOLO11 model and the specified dataset, followed by a summary of the model architecture and training configuration. - Initialization messages for various training components like TensorBoard, Automatic [Mixed Precision](https://www.ultralytics.com/glossary/mixed-precision) (AMP), and dataset preparation. - Finally, the training process starts, with progress updates as the model trains on the specified dataset. For an in-depth understanding of the performance metrics used during training, read [our guide on performance metrics](../guides/yolo-performance-metrics.md). @@ -151,7 +151,7 @@ For a visual walkthrough of what the ClearML Results Page looks like, watch the allowfullscreen>
@@ -34,7 +34,7 @@ Learn how to train a YOLOv8 model with custom data on YouTube with Nicolai. Chec
allowfullscreen>
- Watch: How to Train Ultralytics YOLOv8 models on Your Custom Dataset in Google Colab | Episode 3
+ Watch: How to Train Ultralytics YOLO11 models on Your Custom Dataset in Google Colab | Episode 3
@@ -18,7 +18,7 @@ This Gradio interface provides an easy and interactive way to perform object det
allowfullscreen>
- Watch: Gradio Integration with Ultralytics YOLOv8
+ Watch: Gradio Integration with Ultralytics YOLO11
-- **Optimized Model Efficiency**: Uses pruning and quantization to enhance YOLOv8's efficiency, reducing model size and computational requirements while maintaining [accuracy](https://www.ultralytics.com/glossary/accuracy). +- **Optimized Model Efficiency**: Uses pruning and quantization to enhance YOLO11's efficiency, reducing model size and computational requirements while maintaining [accuracy](https://www.ultralytics.com/glossary/accuracy).
@@ -36,9 +36,9 @@ Before diving into how to deploy YOLOV8 using DeepSparse, let's understand the b - **High Performance on Standard CPUs**: Delivers GPU-like performance on CPUs, providing a more accessible and cost-effective option for various applications. -- **Streamlined Integration and Deployment**: Offers user-friendly tools for easy integration of YOLOv8 into applications, including image and video annotation features. +- **Streamlined Integration and Deployment**: Offers user-friendly tools for easy integration of YOLO11 into applications, including image and video annotation features. -- **Support for Various Model Types**: Compatible with both standard and sparsity-optimized YOLOv8 models, adding deployment flexibility. +- **Support for Various Model Types**: Compatible with both standard and sparsity-optimized YOLO11 models, adding deployment flexibility. - **Cost-Effective and Scalable Solution**: Reduces operational expenses and offers scalable deployment of advanced object detection models. @@ -56,15 +56,15 @@ Neural Magic's Deep Sparse technology is inspired by the human brain's efficienc For more details on how Neural Magic's DeepSparse technology work, check out [their blog post](https://neuralmagic.com/blog/how-neural-magics-deep-sparse-technology-works/). -## Creating A Sparse Version of YOLOv8 Trained on a Custom Dataset +## Creating A Sparse Version of YOLO11 Trained on a Custom Dataset -SparseZoo, an open-source model repository by Neural Magic, offers [a collection of pre-sparsified YOLOv8 model checkpoints](https://sparsezoo.neuralmagic.com/?modelSet=computer_vision&searchModels=yolo). With SparseML, seamlessly integrated with Ultralytics, users can effortlessly fine-tune these sparse checkpoints on their specific datasets using a straightforward command-line interface. +SparseZoo, an open-source model repository by Neural Magic, offers [a collection of pre-sparsified YOLO11 model checkpoints](https://sparsezoo.neuralmagic.com/?modelSet=computer_vision&searchModels=yolo). With SparseML, seamlessly integrated with Ultralytics, users can effortlessly fine-tune these sparse checkpoints on their specific datasets using a straightforward command-line interface. -Checkout [Neural Magic's SparseML YOLOv8 documentation](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov8) for more details. +Checkout [Neural Magic's SparseML YOLO11 documentation](https://github.com/neuralmagic/sparseml/tree/main/integrations/ultralytics-yolov8) for more details. ## Usage: Deploying YOLOV8 using DeepSparse -Deploying YOLOv8 with Neural Magic's DeepSparse involves a few straightforward steps. Before diving into the usage instructions, be sure to check out the range of [YOLOv8 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. Here's how you can get started. +Deploying YOLO11 with Neural Magic's DeepSparse involves a few straightforward steps. Before diving into the usage instructions, be sure to check out the range of [YOLO11 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. Here's how you can get started. ### Step 1: Installation @@ -79,24 +79,24 @@ To install the required packages, run: pip install deepsparse[yolov8] ``` -### Step 2: Exporting YOLOv8 to ONNX Format +### Step 2: Exporting YOLO11 to ONNX Format -DeepSparse Engine requires YOLOv8 models in ONNX format. Exporting your model to this format is essential for compatibility with DeepSparse. Use the following command to export YOLOv8 models: +DeepSparse Engine requires YOLO11 models in ONNX format. Exporting your model to this format is essential for compatibility with DeepSparse. Use the following command to export YOLO11 models: !!! tip "Model Export" === "CLI" ```bash - # Export YOLOv8 model to ONNX format - yolo task=detect mode=export model=yolov8n.pt format=onnx opset=13 + # Export YOLO11 model to ONNX format + yolo task=detect mode=export model=yolo11n.pt format=onnx opset=13 ``` -This command will save the `yolov8n.onnx` model to your disk. +This command will save the `yolo11n.onnx` model to your disk. ### Step 3: Deploying and Running Inferences -With your YOLOv8 model in ONNX format, you can deploy and run inferences using DeepSparse. This can be done easily with their intuitive Python API: +With your YOLO11 model in ONNX format, you can deploy and run inferences using DeepSparse. This can be done easily with their intuitive Python API: !!! tip "Deploying and Running Inferences" @@ -105,8 +105,8 @@ With your YOLOv8 model in ONNX format, you can deploy and run inferences using D ```python from deepsparse import Pipeline - # Specify the path to your YOLOv8 ONNX model - model_path = "path/to/yolov8n.onnx" + # Specify the path to your YOLO11 ONNX model + model_path = "path/to/yolo11n.onnx" # Set up the DeepSparse Pipeline yolo_pipeline = Pipeline.create(task="yolov8", model_path=model_path) @@ -118,7 +118,7 @@ With your YOLOv8 model in ONNX format, you can deploy and run inferences using D ### Step 4: Benchmarking Performance -It's important to check that your YOLOv8 model is performing optimally on DeepSparse. You can benchmark your model's performance to analyze throughput and latency: +It's important to check that your YOLO11 model is performing optimally on DeepSparse. You can benchmark your model's performance to analyze throughput and latency: !!! tip "Benchmarking" @@ -126,12 +126,12 @@ It's important to check that your YOLOv8 model is performing optimally on DeepSp ```bash # Benchmark performance - deepsparse.benchmark model_path="path/to/yolov8n.onnx" --scenario=sync --input_shapes="[1,3,640,640]" + deepsparse.benchmark model_path="path/to/yolo11n.onnx" --scenario=sync --input_shapes="[1,3,640,640]" ``` ### Step 5: Additional Features -DeepSparse provides additional features for practical integration of YOLOv8 in applications, such as image annotation and dataset evaluation. +DeepSparse provides additional features for practical integration of YOLO11 in applications, such as image annotation and dataset evaluation. !!! tip "Additional Features" @@ -139,10 +139,10 @@ DeepSparse provides additional features for practical integration of YOLOv8 in a ```bash # For image annotation - deepsparse.yolov8.annotate --source "path/to/image.jpg" --model_filepath "path/to/yolov8n.onnx" + deepsparse.yolov8.annotate --source "path/to/image.jpg" --model_filepath "path/to/yolo11n.onnx" # For evaluating model performance on a dataset - deepsparse.yolov8.eval --model_path "path/to/yolov8n.onnx" + deepsparse.yolov8.eval --model_path "path/to/yolo11n.onnx" ``` Running the annotate command processes your specified image, detecting objects, and saving the annotated image with bounding boxes and classifications. The annotated image will be stored in an annotation-results folder. This helps provide a visual representation of the model's detection capabilities. @@ -151,61 +151,61 @@ Running the annotate command processes your specified image, detecting objects,
-After running the eval command, you will receive detailed output metrics such as [precision](https://www.ultralytics.com/glossary/precision), [recall](https://www.ultralytics.com/glossary/recall), and mAP (mean Average Precision). This provides a comprehensive view of your model's performance on the dataset. This functionality is particularly useful for fine-tuning and optimizing your YOLOv8 models for specific use cases, ensuring high accuracy and efficiency. +After running the eval command, you will receive detailed output metrics such as [precision](https://www.ultralytics.com/glossary/precision), [recall](https://www.ultralytics.com/glossary/recall), and mAP (mean Average Precision). This provides a comprehensive view of your model's performance on the dataset. This functionality is particularly useful for fine-tuning and optimizing your YOLO11 models for specific use cases, ensuring high accuracy and efficiency. ## Summary -This guide explored integrating Ultralytics' YOLOv8 with Neural Magic's DeepSparse Engine. It highlighted how this integration enhances YOLOv8's performance on CPU platforms, offering GPU-level efficiency and advanced neural network sparsity techniques. +This guide explored integrating Ultralytics' YOLO11 with Neural Magic's DeepSparse Engine. It highlighted how this integration enhances YOLO11's performance on CPU platforms, offering GPU-level efficiency and advanced neural network sparsity techniques. -For more detailed information and advanced usage, visit [Neural Magic's DeepSparse documentation](https://docs.neuralmagic.com/products/deepsparse/). Also, check out Neural Magic's documentation on the integration with YOLOv8 [here](https://github.com/neuralmagic/deepsparse/tree/main/src/deepsparse/yolov8#yolov8-inference-pipelines) and watch a great session on it [here](https://www.youtube.com/watch?v=qtJ7bdt52x8). +For more detailed information and advanced usage, visit [Neural Magic's DeepSparse documentation](https://docs.neuralmagic.com/products/deepsparse/). Also, check out Neural Magic's documentation on the integration with YOLO11 [here](https://github.com/neuralmagic/deepsparse/tree/main/src/deepsparse/yolov8#yolov8-inference-pipelines) and watch a great session on it [here](https://www.youtube.com/watch?v=qtJ7bdt52x8). -Additionally, for a broader understanding of various YOLOv8 integrations, visit the [Ultralytics integration guide page](../integrations/index.md), where you can discover a range of other exciting integration possibilities. +Additionally, for a broader understanding of various YOLO11 integrations, visit the [Ultralytics integration guide page](../integrations/index.md), where you can discover a range of other exciting integration possibilities. ## FAQ -### What is Neural Magic's DeepSparse Engine and how does it optimize YOLOv8 performance? +### What is Neural Magic's DeepSparse Engine and how does it optimize YOLO11 performance? -Neural Magic's DeepSparse Engine is an inference runtime designed to optimize the execution of neural networks on CPUs through advanced techniques such as sparsity, pruning, and quantization. By integrating DeepSparse with YOLOv8, you can achieve GPU-like performance on standard CPUs, significantly enhancing inference speed, model efficiency, and overall performance while maintaining accuracy. For more details, check out the [Neural Magic's DeepSparse section](#neural-magics-deepsparse). +Neural Magic's DeepSparse Engine is an inference runtime designed to optimize the execution of neural networks on CPUs through advanced techniques such as sparsity, pruning, and quantization. By integrating DeepSparse with YOLO11, you can achieve GPU-like performance on standard CPUs, significantly enhancing inference speed, model efficiency, and overall performance while maintaining accuracy. For more details, check out the [Neural Magic's DeepSparse section](#neural-magics-deepsparse). -### How can I install the needed packages to deploy YOLOv8 using Neural Magic's DeepSparse? +### How can I install the needed packages to deploy YOLO11 using Neural Magic's DeepSparse? -Installing the required packages for deploying YOLOv8 with Neural Magic's DeepSparse is straightforward. You can easily install them using the CLI. Here's the command you need to run: +Installing the required packages for deploying YOLO11 with Neural Magic's DeepSparse is straightforward. You can easily install them using the CLI. Here's the command you need to run: ```bash pip install deepsparse[yolov8] ``` -Once installed, follow the steps provided in the [Installation section](#step-1-installation) to set up your environment and start using DeepSparse with YOLOv8. +Once installed, follow the steps provided in the [Installation section](#step-1-installation) to set up your environment and start using DeepSparse with YOLO11. -### How do I convert YOLOv8 models to ONNX format for use with DeepSparse? +### How do I convert YOLO11 models to ONNX format for use with DeepSparse? -To convert YOLOv8 models to the ONNX format, which is required for compatibility with DeepSparse, you can use the following CLI command: +To convert YOLO11 models to the ONNX format, which is required for compatibility with DeepSparse, you can use the following CLI command: ```bash -yolo task=detect mode=export model=yolov8n.pt format=onnx opset=13 +yolo task=detect mode=export model=yolo11n.pt format=onnx opset=13 ``` -This command will export your YOLOv8 model (`yolov8n.pt`) to a format (`yolov8n.onnx`) that can be utilized by the DeepSparse Engine. More information about model export can be found in the [Model Export section](#step-2-exporting-yolov8-to-onnx-format). +This command will export your YOLO11 model (`yolo11n.pt`) to a format (`yolo11n.onnx`) that can be utilized by the DeepSparse Engine. More information about model export can be found in the [Model Export section](#step-2-exporting-yolo11-to-onnx-format). -### How do I benchmark YOLOv8 performance on the DeepSparse Engine? +### How do I benchmark YOLO11 performance on the DeepSparse Engine? -Benchmarking YOLOv8 performance on DeepSparse helps you analyze throughput and latency to ensure your model is optimized. You can use the following CLI command to run a benchmark: +Benchmarking YOLO11 performance on DeepSparse helps you analyze throughput and latency to ensure your model is optimized. You can use the following CLI command to run a benchmark: ```bash -deepsparse.benchmark model_path="path/to/yolov8n.onnx" --scenario=sync --input_shapes="[1,3,640,640]" +deepsparse.benchmark model_path="path/to/yolo11n.onnx" --scenario=sync --input_shapes="[1,3,640,640]" ``` This command will provide you with vital performance metrics. For more details, see the [Benchmarking Performance section](#step-4-benchmarking-performance). -### Why should I use Neural Magic's DeepSparse with YOLOv8 for object detection tasks? +### Why should I use Neural Magic's DeepSparse with YOLO11 for object detection tasks? -Integrating Neural Magic's DeepSparse with YOLOv8 offers several benefits: +Integrating Neural Magic's DeepSparse with YOLO11 offers several benefits: -- **Enhanced Inference Speed:** Achieves up to 525 FPS, significantly speeding up YOLOv8's capabilities. +- **Enhanced Inference Speed:** Achieves up to 525 FPS, significantly speeding up YOLO11's capabilities. - **Optimized Model Efficiency:** Uses sparsity, pruning, and quantization techniques to reduce model size and computational needs while maintaining accuracy. - **High Performance on Standard CPUs:** Offers GPU-like performance on cost-effective CPU hardware. - **Streamlined Integration:** User-friendly tools for easy deployment and integration. -- **Flexibility:** Supports both standard and sparsity-optimized YOLOv8 models. +- **Flexibility:** Supports both standard and sparsity-optimized YOLO11 models. - **Cost-Effective:** Reduces operational expenses through efficient resource utilization. -For a deeper dive into these advantages, visit the [Benefits of Integrating Neural Magic's DeepSparse with YOLOv8 section](#benefits-of-integrating-neural-magics-deepsparse-with-yolov8). +For a deeper dive into these advantages, visit the [Benefits of Integrating Neural Magic's DeepSparse with YOLO11 section](#benefits-of-integrating-neural-magics-deepsparse-with-yolo11). diff --git a/docs/en/integrations/onnx.md b/docs/en/integrations/onnx.md index 3bb372ac2a..fbff328d02 100644 --- a/docs/en/integrations/onnx.md +++ b/docs/en/integrations/onnx.md @@ -1,14 +1,14 @@ --- comments: true -description: Learn how to export YOLOv8 models to ONNX format for flexible deployment across various platforms with enhanced performance. -keywords: YOLOv8, ONNX, model export, Ultralytics, ONNX Runtime, machine learning, model deployment, computer vision, deep learning +description: Learn how to export YOLO11 models to ONNX format for flexible deployment across various platforms with enhanced performance. +keywords: YOLO11, ONNX, model export, Ultralytics, ONNX Runtime, machine learning, model deployment, computer vision, deep learning --- -# ONNX Export for YOLOv8 Models +# ONNX Export for YOLO11 Models Often, when deploying [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models, you'll need a model format that's both flexible and compatible with multiple platforms. -Exporting [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models to ONNX format streamlines deployment and ensures optimal performance across various environments. This guide will show you how to easily convert your YOLOv8 models to ONNX and enhance their scalability and effectiveness in real-world applications. +Exporting [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models to ONNX format streamlines deployment and ensures optimal performance across various environments. This guide will show you how to easily convert your YOLO11 models to ONNX and enhance their scalability and effectiveness in real-world applications. ## ONNX and ONNX Runtime @@ -44,7 +44,7 @@ The ability of ONNX to handle various formats can be attributed to the following ## Common Usage of ONNX -Before we jump into how to export YOLOv8 models to the ONNX format, let's take a look at where ONNX models are usually used. +Before we jump into how to export YOLO11 models to the ONNX format, let's take a look at where ONNX models are usually used. ### CPU Deployment @@ -60,9 +60,9 @@ While ONNX models are commonly used on CPUs, they can also be deployed on the fo - **Web Browsers**: ONNX can run directly in web browsers, powering interactive and dynamic web-based AI applications. -## Exporting YOLOv8 Models to ONNX +## Exporting YOLO11 Models to ONNX -You can expand model compatibility and deployment flexibility by converting YOLOv8 models to ONNX format. +You can expand model compatibility and deployment flexibility by converting YOLO11 models to ONNX format. ### Installation @@ -73,15 +73,15 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [YOLOv8 Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [YOLO11 Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, be sure to check out the range of [YOLOv8 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. +Before diving into the usage instructions, be sure to check out the range of [YOLO11 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. !!! example "Usage" @@ -90,14 +90,14 @@ Before diving into the usage instructions, be sure to check out the range of [YO ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to ONNX format - model.export(format="onnx") # creates 'yolov8n.onnx' + model.export(format="onnx") # creates 'yolo11n.onnx' # Load the exported ONNX model - onnx_model = YOLO("yolov8n.onnx") + onnx_model = YOLO("yolo11n.onnx") # Run inference results = onnx_model("https://ultralytics.com/images/bus.jpg") @@ -106,18 +106,18 @@ Before diving into the usage instructions, be sure to check out the range of [YO === "CLI" ```bash - # Export a YOLOv8n PyTorch model to ONNX format - yolo export model=yolov8n.pt format=onnx # creates 'yolov8n.onnx' + # Export a YOLO11n PyTorch model to ONNX format + yolo export model=yolo11n.pt format=onnx # creates 'yolo11n.onnx' # Run inference with the exported model - yolo predict model=yolov8n.onnx source='https://ultralytics.com/images/bus.jpg' + yolo predict model=yolo11n.onnx source='https://ultralytics.com/images/bus.jpg' ``` For more details about the export process, visit the [Ultralytics documentation page on exporting](../modes/export.md). -## Deploying Exported YOLOv8 ONNX Models +## Deploying Exported YOLO11 ONNX Models -Once you've successfully exported your Ultralytics YOLOv8 models to ONNX format, the next step is deploying these models in various environments. For detailed instructions on deploying your ONNX models, take a look at the following resources: +Once you've successfully exported your Ultralytics YOLO11 models to ONNX format, the next step is deploying these models in various environments. For detailed instructions on deploying your ONNX models, take a look at the following resources: - **[ONNX Runtime Python API Documentation](https://onnxruntime.ai/docs/api/python/api_summary.html)**: This guide provides essential information for loading and running ONNX models using ONNX Runtime. @@ -127,17 +127,17 @@ Once you've successfully exported your Ultralytics YOLOv8 models to ONNX format, ## Summary -In this guide, you've learned how to export Ultralytics YOLOv8 models to ONNX format to increase their interoperability and performance across various platforms. You were also introduced to the ONNX Runtime and ONNX deployment options. +In this guide, you've learned how to export Ultralytics YOLO11 models to ONNX format to increase their interoperability and performance across various platforms. You were also introduced to the ONNX Runtime and ONNX deployment options. For further details on usage, visit the [ONNX official documentation](https://onnx.ai/onnx/intro/). -Also, if you'd like to know more about other Ultralytics YOLOv8 integrations, visit our [integration guide page](../integrations/index.md). You'll find plenty of useful resources and insights there. +Also, if you'd like to know more about other Ultralytics YOLO11 integrations, visit our [integration guide page](../integrations/index.md). You'll find plenty of useful resources and insights there. ## FAQ -### How do I export YOLOv8 models to ONNX format using Ultralytics? +### How do I export YOLO11 models to ONNX format using Ultralytics? -To export your YOLOv8 models to ONNX format using Ultralytics, follow these steps: +To export your YOLO11 models to ONNX format using Ultralytics, follow these steps: !!! example "Usage" @@ -146,14 +146,14 @@ To export your YOLOv8 models to ONNX format using Ultralytics, follow these step ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to ONNX format - model.export(format="onnx") # creates 'yolov8n.onnx' + model.export(format="onnx") # creates 'yolo11n.onnx' # Load the exported ONNX model - onnx_model = YOLO("yolov8n.onnx") + onnx_model = YOLO("yolo11n.onnx") # Run inference results = onnx_model("https://ultralytics.com/images/bus.jpg") @@ -162,18 +162,18 @@ To export your YOLOv8 models to ONNX format using Ultralytics, follow these step === "CLI" ```bash - # Export a YOLOv8n PyTorch model to ONNX format - yolo export model=yolov8n.pt format=onnx # creates 'yolov8n.onnx' + # Export a YOLO11n PyTorch model to ONNX format + yolo export model=yolo11n.pt format=onnx # creates 'yolo11n.onnx' # Run inference with the exported model - yolo predict model=yolov8n.onnx source='https://ultralytics.com/images/bus.jpg' + yolo predict model=yolo11n.onnx source='https://ultralytics.com/images/bus.jpg' ``` For more details, visit the [export documentation](../modes/export.md). -### What are the advantages of using ONNX Runtime for deploying YOLOv8 models? +### What are the advantages of using ONNX Runtime for deploying YOLO11 models? -Using ONNX Runtime for deploying YOLOv8 models offers several advantages: +Using ONNX Runtime for deploying YOLO11 models offers several advantages: - **Cross-platform compatibility**: ONNX Runtime supports various platforms, such as Windows, macOS, and Linux, ensuring your models run smoothly across different environments. - **Hardware acceleration**: ONNX Runtime can leverage hardware-specific optimizations for CPUs, GPUs, and dedicated accelerators, providing high-performance inference. @@ -181,9 +181,9 @@ Using ONNX Runtime for deploying YOLOv8 models offers several advantages: Learn more by checking the [ONNX Runtime documentation](https://onnxruntime.ai/docs/api/python/api_summary.html). -### What deployment options are available for YOLOv8 models exported to ONNX? +### What deployment options are available for YOLO11 models exported to ONNX? -YOLOv8 models exported to ONNX can be deployed on various platforms including: +YOLO11 models exported to ONNX can be deployed on various platforms including: - **CPUs**: Utilizing ONNX Runtime for optimized CPU inference. - **GPUs**: Leveraging NVIDIA CUDA for high-performance GPU acceleration. @@ -192,19 +192,19 @@ YOLOv8 models exported to ONNX can be deployed on various platforms including: For more information, explore our guide on [model deployment options](../guides/model-deployment-options.md). -### Why should I use ONNX format for Ultralytics YOLOv8 models? +### Why should I use ONNX format for Ultralytics YOLO11 models? -Using ONNX format for Ultralytics YOLOv8 models provides numerous benefits: +Using ONNX format for Ultralytics YOLO11 models provides numerous benefits: - **Interoperability**: ONNX allows models to be transferred between different machine learning frameworks seamlessly. - **Performance Optimization**: ONNX Runtime can enhance model performance by utilizing hardware-specific optimizations. - **Flexibility**: ONNX supports various deployment environments, enabling you to use the same model on different platforms without modification. -Refer to the comprehensive guide on [exporting YOLOv8 models to ONNX](https://www.ultralytics.com/blog/export-and-optimize-a-yolov8-model-for-inference-on-openvino). +Refer to the comprehensive guide on [exporting YOLO11 models to ONNX](https://www.ultralytics.com/blog/export-and-optimize-a-yolov8-model-for-inference-on-openvino). -### How can I troubleshoot issues when exporting YOLOv8 models to ONNX? +### How can I troubleshoot issues when exporting YOLO11 models to ONNX? -When exporting YOLOv8 models to ONNX, you might encounter common issues such as mismatched dependencies or unsupported operations. To troubleshoot these problems: +When exporting YOLO11 models to ONNX, you might encounter common issues such as mismatched dependencies or unsupported operations. To troubleshoot these problems: 1. Verify that you have the correct version of required dependencies installed. 2. Check the official [ONNX documentation](https://onnx.ai/onnx/intro/) for supported operators and features. diff --git a/docs/en/integrations/paddlepaddle.md b/docs/en/integrations/paddlepaddle.md index 62092df9f5..c88bacde89 100644 --- a/docs/en/integrations/paddlepaddle.md +++ b/docs/en/integrations/paddlepaddle.md @@ -1,12 +1,12 @@ --- comments: true -description: Learn how to export YOLOv8 models to PaddlePaddle format for enhanced performance, flexibility, and deployment across various platforms and devices. -keywords: YOLOv8, PaddlePaddle, export models, computer vision, deep learning, model deployment, performance optimization +description: Learn how to export YOLO11 models to PaddlePaddle format for enhanced performance, flexibility, and deployment across various platforms and devices. +keywords: YOLO11, PaddlePaddle, export models, computer vision, deep learning, model deployment, performance optimization --- -# How to Export to PaddlePaddle Format from YOLOv8 Models +# How to Export to PaddlePaddle Format from YOLO11 Models -Bridging the gap between developing and deploying [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models in real-world scenarios with varying conditions can be difficult. PaddlePaddle makes this process easier with its focus on flexibility, performance, and its capability for parallel processing in distributed environments. This means you can use your YOLOv8 computer vision models on a wide variety of devices and platforms, from smartphones to cloud-based servers. +Bridging the gap between developing and deploying [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models in real-world scenarios with varying conditions can be difficult. PaddlePaddle makes this process easier with its focus on flexibility, performance, and its capability for parallel processing in distributed environments. This means you can use your YOLO11 computer vision models on a wide variety of devices and platforms, from smartphones to cloud-based servers.
@@ -16,10 +16,10 @@ Bridging the gap between developing and deploying [computer vision](https://www.
allowfullscreen>
- Watch: How to Export Ultralytics YOLOv8 Models to PaddlePaddle Format | Key Features of PaddlePaddle Format
+ Watch: How to Export Ultralytics YOLO11 Models to PaddlePaddle Format | Key Features of PaddlePaddle Format
@@ -46,14 +46,14 @@ Explore more capabilities of YOLOv8 and Paperspace Gradient in a discussion with
allowfullscreen>
- Watch: Ultralytics Live Session 7: It's All About the Environment: Optimizing YOLOv8 Training With Gradient
+ Watch: Ultralytics Live Session 7: It's All About the Environment: Optimizing YOLO11 Training With Gradient
-Universe also has a page that aggregates all [public fine-tuned YOLOv8 models uploaded to Roboflow](https://universe.roboflow.com/search?q=model%3Ayolov8&ref=ultralytics). You can use this page to explore pre-trained models you can use for testing or [for automated data labeling](https://docs.roboflow.com/annotate/use-roboflow-annotate/model-assisted-labeling?ref=ultralytics) or to prototype with [Roboflow inference](https://github.com/roboflow/inference?ref=ultralytics). +Universe also has a page that aggregates all [public fine-tuned YOLO11 models uploaded to Roboflow](https://universe.roboflow.com/search?q=model%3Ayolov8&ref=ultralytics). You can use this page to explore pre-trained models you can use for testing or [for automated data labeling](https://docs.roboflow.com/annotate/use-roboflow-annotate/model-assisted-labeling?ref=ultralytics) or to prototype with [Roboflow inference](https://github.com/roboflow/inference?ref=ultralytics). If you want to gather images yourself, try [Collect](https://github.com/roboflow/roboflow-collect), an open source project that allows you to automatically gather images using a webcam on the edge. You can use text or image prompts with Collect to instruct what data should be collected, allowing you to capture only the useful data you need to build your vision model. -## Upload, Convert and Label Data for YOLOv8 Format +## Upload, Convert and Label Data for YOLO11 Format [Roboflow Annotate](https://docs.roboflow.com/annotate/use-roboflow-annotate?ref=ultralytics) is an online annotation tool for use in labeling images for [object detection](https://www.ultralytics.com/glossary/object-detection), classification, and segmentation. -To label data for a YOLOv8 object detection, [instance segmentation](https://www.ultralytics.com/glossary/instance-segmentation), or classification model, first create a project in Roboflow. +To label data for a YOLO11 object detection, [instance segmentation](https://www.ultralytics.com/glossary/instance-segmentation), or classification model, first create a project in Roboflow.
@@ -95,7 +95,7 @@ You can also add tags to images from the Tags panel in the sidebar. You can appl
-Models hosted on Roboflow can be used with Label Assist, an automated annotation tool that uses your YOLOv8 model to recommend annotations. To use Label Assist, first upload a YOLOv8 model to Roboflow (see instructions later in the guide). Then, click the magic wand icon in the left sidebar and select your model for use in Label Assist. +Models hosted on Roboflow can be used with Label Assist, an automated annotation tool that uses your YOLO11 model to recommend annotations. To use Label Assist, first upload a YOLO11 model to Roboflow (see instructions later in the guide). Then, click the magic wand icon in the left sidebar and select your model for use in Label Assist. Choose a model, then click "Continue" to enable Label Assist: @@ -109,7 +109,7 @@ When you open new images for annotation, Label Assist will trigger and recommend -## Dataset Management for YOLOv8 +## Dataset Management for YOLO11 Roboflow provides a suite of tools for understanding computer vision datasets. @@ -157,13 +157,13 @@ When your dataset version has been generated, you can export your data into a ra -You are now ready to train YOLOv8 on a custom dataset. Follow this [written guide](https://blog.roboflow.com/how-to-train-yolov8-on-a-custom-dataset/?ref=ultralytics) and [YouTube video](https://www.youtube.com/watch?v=wuZtUMEiKWY) for step-by-step instructions or refer to the [Ultralytics documentation](../modes/train.md). +You are now ready to train YOLO11 on a custom dataset. Follow this [written guide](https://blog.roboflow.com/how-to-train-yolov8-on-a-custom-dataset/?ref=ultralytics) and [YouTube video](https://www.youtube.com/watch?v=wuZtUMEiKWY) for step-by-step instructions or refer to the [Ultralytics documentation](../modes/train.md). -## Upload Custom YOLOv8 Model Weights for Testing and Deployment +## Upload Custom YOLO11 Model Weights for Testing and Deployment Roboflow offers an infinitely scalable API for deployed models and SDKs for use with NVIDIA Jetsons, Luxonis OAKs, Raspberry Pis, GPU-based devices, and more. -You can deploy YOLOv8 models by uploading YOLOv8 weights to Roboflow. You can do this in a few lines of Python code. Create a new Python file and add the following code: +You can deploy YOLO11 models by uploading YOLO11 weights to Roboflow. You can do this in a few lines of Python code. Create a new Python file and add the following code: ```python import roboflow # install with 'pip install roboflow' @@ -190,7 +190,7 @@ To test your model and find deployment instructions for supported SDKs, go to th You can also use your uploaded model as a [labeling assistant](https://docs.roboflow.com/annotate/use-roboflow-annotate/model-assisted-labeling?ref=ultralytics). This feature uses your trained model to recommend annotations on images uploaded to Roboflow. -## How to Evaluate YOLOv8 Models +## How to Evaluate YOLO11 Models Roboflow provides a range of features for use in evaluating models. @@ -224,17 +224,17 @@ You can use Vector Analysis to: ## Learning Resources -Want to learn more about using Roboflow for creating YOLOv8 models? The following resources may be helpful in your work. +Want to learn more about using Roboflow for creating YOLO11 models? The following resources may be helpful in your work. -- [Train YOLOv8 on a Custom Dataset](https://github.com/roboflow/notebooks/blob/main/notebooks/train-yolov8-object-detection-on-custom-dataset.ipynb): Follow our interactive notebook that shows you how to train a YOLOv8 model on a custom dataset. -- [Autodistill](https://docs.autodistill.com/): Use large foundation vision models to label data for specific models. You can label images for use in training YOLOv8 classification, detection, and segmentation models with Autodistill. +- [Train YOLO11 on a Custom Dataset](https://github.com/roboflow/notebooks/blob/main/notebooks/train-yolov8-object-detection-on-custom-dataset.ipynb): Follow our interactive notebook that shows you how to train a YOLO11 model on a custom dataset. +- [Autodistill](https://docs.autodistill.com/): Use large foundation vision models to label data for specific models. You can label images for use in training YOLO11 classification, detection, and segmentation models with Autodistill. - [Supervision](https://supervision.roboflow.com/?ref=ultralytics): A Python package with helpful utilities for use in working with computer vision models. You can use supervision to filter detections, compute confusion matrices, and more, all in a few lines of Python code. -- [Roboflow Blog](https://blog.roboflow.com/?ref=ultralytics): The Roboflow Blog features over 500 articles on computer vision, covering topics from how to train a YOLOv8 model to annotation best practices. -- [Roboflow YouTube channel](https://www.youtube.com/@Roboflow): Browse dozens of in-depth computer vision guides on our YouTube channel, covering topics from training YOLOv8 models to automated image labeling. +- [Roboflow Blog](https://blog.roboflow.com/?ref=ultralytics): The Roboflow Blog features over 500 articles on computer vision, covering topics from how to train a YOLO11 model to annotation best practices. +- [Roboflow YouTube channel](https://www.youtube.com/@Roboflow): Browse dozens of in-depth computer vision guides on our YouTube channel, covering topics from training YOLO11 models to automated image labeling. ## Project Showcase -Below are a few of the many pieces of feedback we have received for using YOLOv8 and Roboflow together to create computer vision models. +Below are a few of the many pieces of feedback we have received for using YOLO11 and Roboflow together to create computer vision models.@@ -244,26 +244,26 @@ Below are a few of the many pieces of feedback we have received for using YOLOv8 ## FAQ -### How do I label data for YOLOv8 models using Roboflow? +### How do I label data for YOLO11 models using Roboflow? -Labeling data for YOLOv8 models using Roboflow is straightforward with Roboflow Annotate. First, create a project on Roboflow and upload your images. After uploading, select the batch of images and click "Start Annotating." You can use the `B` key for bounding boxes or the `P` key for polygons. For faster annotation, use the SAM-based label assistant by clicking the cursor icon in the sidebar. Detailed steps can be found [here](#upload-convert-and-label-data-for-yolov8-format). +Labeling data for YOLO11 models using Roboflow is straightforward with Roboflow Annotate. First, create a project on Roboflow and upload your images. After uploading, select the batch of images and click "Start Annotating." You can use the `B` key for bounding boxes or the `P` key for polygons. For faster annotation, use the SAM-based label assistant by clicking the cursor icon in the sidebar. Detailed steps can be found [here](#upload-convert-and-label-data-for-yolo11-format). -### What services does Roboflow offer for collecting YOLOv8 [training data](https://www.ultralytics.com/glossary/training-data)? +### What services does Roboflow offer for collecting YOLO11 [training data](https://www.ultralytics.com/glossary/training-data)? -Roboflow provides two key services for collecting YOLOv8 training data: [Universe](https://universe.roboflow.com/?ref=ultralytics) and [Collect](https://github.com/roboflow/roboflow-collect?ref=ultralytics). Universe offers access to over 250,000 vision datasets, while Collect helps you gather images using a webcam and automated prompts. +Roboflow provides two key services for collecting YOLO11 training data: [Universe](https://universe.roboflow.com/?ref=ultralytics) and [Collect](https://github.com/roboflow/roboflow-collect?ref=ultralytics). Universe offers access to over 250,000 vision datasets, while Collect helps you gather images using a webcam and automated prompts. -### How can I manage and analyze my YOLOv8 dataset using Roboflow? +### How can I manage and analyze my YOLO11 dataset using Roboflow? -Roboflow offers robust dataset management tools, including dataset search, tagging, and Health Check. Use the search feature to find images based on text descriptions or tags. Health Check provides insights into dataset quality, showing class balance, image sizes, and annotation heatmaps. This helps optimize dataset performance before training YOLOv8 models. Detailed information can be found [here](#dataset-management-for-yolov8). +Roboflow offers robust dataset management tools, including dataset search, tagging, and Health Check. Use the search feature to find images based on text descriptions or tags. Health Check provides insights into dataset quality, showing class balance, image sizes, and annotation heatmaps. This helps optimize dataset performance before training YOLO11 models. Detailed information can be found [here](#dataset-management-for-yolo11). -### How do I export my YOLOv8 dataset from Roboflow? +### How do I export my YOLO11 dataset from Roboflow? -To export your YOLOv8 dataset from Roboflow, you need to create a dataset version. Click "Versions" in the sidebar, then "Create New Version" and apply any desired augmentations. Once the version is generated, click "Export Dataset" and choose the YOLOv8 format. Follow this process [here](#export-data-in-40-formats-for-model-training). +To export your YOLO11 dataset from Roboflow, you need to create a dataset version. Click "Versions" in the sidebar, then "Create New Version" and apply any desired augmentations. Once the version is generated, click "Export Dataset" and choose the YOLO11 format. Follow this process [here](#export-data-in-40-formats-for-model-training). -### How can I integrate and deploy YOLOv8 models with Roboflow? +### How can I integrate and deploy YOLO11 models with Roboflow? -Integrate and deploy YOLOv8 models on Roboflow by uploading your YOLOv8 weights through a few lines of Python code. Use the provided script to authenticate and upload your model, which will create an API for deployment. For details on the script and further instructions, see [this section](#upload-custom-yolov8-model-weights-for-testing-and-deployment). +Integrate and deploy YOLO11 models on Roboflow by uploading your YOLO11 weights through a few lines of Python code. Use the provided script to authenticate and upload your model, which will create an API for deployment. For details on the script and further instructions, see [this section](#upload-custom-yolo11-model-weights-for-testing-and-deployment). -### What tools does Roboflow provide for evaluating YOLOv8 models? +### What tools does Roboflow provide for evaluating YOLO11 models? -Roboflow offers model evaluation tools, including a confusion matrix and vector analysis plots. Access these tools from the "View Detailed Evaluation" button on your model page. These features help identify model performance issues and find areas for improvement. For more information, refer to [this section](#how-to-evaluate-yolov8-models). +Roboflow offers model evaluation tools, including a confusion matrix and vector analysis plots. Access these tools from the "View Detailed Evaluation" button on your model page. These features help identify model performance issues and find areas for improvement. For more information, refer to [this section](#how-to-evaluate-yolo11-models). diff --git a/docs/en/integrations/tensorboard.md b/docs/en/integrations/tensorboard.md index d563aca12b..c6cad0d50b 100644 --- a/docs/en/integrations/tensorboard.md +++ b/docs/en/integrations/tensorboard.md @@ -1,14 +1,14 @@ --- comments: true -description: Learn how to integrate YOLOv8 with TensorBoard for real-time visual insights into your model's training metrics, performance graphs, and debugging workflows. -keywords: YOLOv8, TensorBoard, model training, visualization, machine learning, deep learning, Ultralytics, training metrics, performance analysis +description: Learn how to integrate YOLO11 with TensorBoard for real-time visual insights into your model's training metrics, performance graphs, and debugging workflows. +keywords: YOLO11, TensorBoard, model training, visualization, machine learning, deep learning, Ultralytics, training metrics, performance analysis --- -# Gain Visual Insights with YOLOv8's Integration with TensorBoard +# Gain Visual Insights with YOLO11's Integration with TensorBoard -Understanding and fine-tuning [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models like [Ultralytics' YOLOv8](https://www.ultralytics.com/) becomes more straightforward when you take a closer look at their training processes. Model training visualization helps with getting insights into the model's learning patterns, performance metrics, and overall behavior. YOLOv8's integration with TensorBoard makes this process of visualization and analysis easier and enables more efficient and informed adjustments to the model. +Understanding and fine-tuning [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models like [Ultralytics' YOLO11](https://www.ultralytics.com/) becomes more straightforward when you take a closer look at their training processes. Model training visualization helps with getting insights into the model's learning patterns, performance metrics, and overall behavior. YOLO11's integration with TensorBoard makes this process of visualization and analysis easier and enables more efficient and informed adjustments to the model. -This guide covers how to use TensorBoard with YOLOv8. You'll learn about various visualizations, from tracking metrics to analyzing model graphs. These tools will help you understand your YOLOv8 model's performance better. +This guide covers how to use TensorBoard with YOLO11. You'll learn about various visualizations, from tracking metrics to analyzing model graphs. These tools will help you understand your YOLO11 model's performance better. ## TensorBoard @@ -18,9 +18,9 @@ This guide covers how to use TensorBoard with YOLOv8. You'll learn about various [TensorBoard](https://www.tensorflow.org/tensorboard), [TensorFlow](https://www.ultralytics.com/glossary/tensorflow)'s visualization toolkit, is essential for [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) experimentation. TensorBoard features a range of visualization tools, crucial for monitoring machine learning models. These tools include tracking key metrics like loss and accuracy, visualizing model graphs, and viewing histograms of weights and biases over time. It also provides capabilities for projecting [embeddings](https://www.ultralytics.com/glossary/embeddings) to lower-dimensional spaces and displaying multimedia data. -## YOLOv8 Training with TensorBoard +## YOLO11 Training with TensorBoard -Using TensorBoard while training YOLOv8 models is straightforward and offers significant benefits. +Using TensorBoard while training YOLO11 models is straightforward and offers significant benefits. ## Installation @@ -31,13 +31,13 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 and Tensorboard + # Install the required package for YOLO11 and Tensorboard pip install ultralytics ``` -TensorBoard is conveniently pre-installed with YOLOv8, eliminating the need for additional setup for visualization purposes. +TensorBoard is conveniently pre-installed with YOLO11, eliminating the need for additional setup for visualization purposes. -For detailed instructions and best practices related to the installation process, be sure to check our [YOLOv8 Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, be sure to check our [YOLO11 Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ## Configuring TensorBoard for Google Colab @@ -54,7 +54,7 @@ When using Google Colab, it's important to set up TensorBoard before starting yo ## Usage -Before diving into the usage instructions, be sure to check out the range of [YOLOv8 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. +Before diving into the usage instructions, be sure to check out the range of [YOLO11 models offered by Ultralytics](../models/index.md). This will help you choose the most appropriate model for your project requirements. !!! example "Usage" @@ -64,7 +64,7 @@ Before diving into the usage instructions, be sure to check out the range of [YO from ultralytics import YOLO # Load a pre-trained model - model = YOLO("yolov8n.pt") + model = YOLO("yolo11n.pt") # Train the model results = model.train(data="coco8.yaml", epochs=100, imgsz=640) @@ -76,17 +76,17 @@ Upon running the usage code snippet above, you can expect the following output: TensorBoard: Start with 'tensorboard --logdir path_to_your_tensorboard_logs', view at http://localhost:6006/ ``` -This output indicates that TensorBoard is now actively monitoring your YOLOv8 training session. You can access the TensorBoard dashboard by visiting the provided URL (http://localhost:6006/) to view real-time training metrics and model performance. For users working in Google Colab, the TensorBoard will be displayed in the same cell where you executed the TensorBoard configuration commands. +This output indicates that TensorBoard is now actively monitoring your YOLO11 training session. You can access the TensorBoard dashboard by visiting the provided URL (http://localhost:6006/) to view real-time training metrics and model performance. For users working in Google Colab, the TensorBoard will be displayed in the same cell where you executed the TensorBoard configuration commands. -For more information related to the model training process, be sure to check our [YOLOv8 Model Training guide](../modes/train.md). If you are interested in learning more about logging, checkpoints, plotting, and file management, read our [usage guide on configuration](../usage/cfg.md). +For more information related to the model training process, be sure to check our [YOLO11 Model Training guide](../modes/train.md). If you are interested in learning more about logging, checkpoints, plotting, and file management, read our [usage guide on configuration](../usage/cfg.md). -## Understanding Your TensorBoard for YOLOv8 Training +## Understanding Your TensorBoard for YOLO11 Training -Now, let's focus on understanding the various features and components of TensorBoard in the context of YOLOv8 training. The three key sections of the TensorBoard are Time Series, Scalars, and Graphs. +Now, let's focus on understanding the various features and components of TensorBoard in the context of YOLO11 training. The three key sections of the TensorBoard are Time Series, Scalars, and Graphs. ### Time Series -The Time Series feature in the TensorBoard offers a dynamic and detailed perspective of various training metrics over time for YOLOv8 models. It focuses on the progression and trends of metrics across training epochs. Here's an example of what you can expect to see. +The Time Series feature in the TensorBoard offers a dynamic and detailed perspective of various training metrics over time for YOLO11 models. It focuses on the progression and trends of metrics across training epochs. Here's an example of what you can expect to see. ![image](https://github.com/ultralytics/docs/releases/download/0/time-series-tensorboard-yolov8.avif) @@ -100,13 +100,13 @@ The Time Series feature in the TensorBoard offers a dynamic and detailed perspec - **In-Depth Analysis**: Time Series provides an in-depth analysis of each metric. For instance, different learning rate segments are shown, offering insights into how adjustments in learning rate impact the model's learning curve. -#### Importance of Time Series in YOLOv8 Training +#### Importance of Time Series in YOLO11 Training -The Time Series section is essential for a thorough analysis of the YOLOv8 model's training progress. It lets you track the metrics in real time to promptly identify and solve issues. It also offers a detailed view of each metrics progression, which is crucial for fine-tuning the model and enhancing its performance. +The Time Series section is essential for a thorough analysis of the YOLO11 model's training progress. It lets you track the metrics in real time to promptly identify and solve issues. It also offers a detailed view of each metrics progression, which is crucial for fine-tuning the model and enhancing its performance. ### Scalars -Scalars in the TensorBoard are crucial for plotting and analyzing simple metrics like loss and accuracy during the training of YOLOv8 models. They offer a clear and concise view of how these metrics evolve with each training [epoch](https://www.ultralytics.com/glossary/epoch), providing insights into the model's learning effectiveness and stability. Here's an example of what you can expect to see. +Scalars in the TensorBoard are crucial for plotting and analyzing simple metrics like loss and accuracy during the training of YOLO11 models. They offer a clear and concise view of how these metrics evolve with each training [epoch](https://www.ultralytics.com/glossary/epoch), providing insights into the model's learning effectiveness and stability. Here's an example of what you can expect to see. ![image](https://github.com/ultralytics/docs/releases/download/0/scalars-metrics-tensorboard.avif) @@ -130,7 +130,7 @@ Scalars in the TensorBoard are crucial for plotting and analyzing simple metrics #### Importance of Monitoring Scalars -Observing scalar metrics is crucial for fine-tuning the YOLOv8 model. Variations in these metrics, such as spikes or irregular patterns in loss graphs, can highlight potential issues such as [overfitting](https://www.ultralytics.com/glossary/overfitting), [underfitting](https://www.ultralytics.com/glossary/underfitting), or inappropriate learning rate settings. By closely monitoring these scalars, you can make informed decisions to optimize the training process, ensuring that the model learns effectively and achieves the desired performance. +Observing scalar metrics is crucial for fine-tuning the YOLO11 model. Variations in these metrics, such as spikes or irregular patterns in loss graphs, can highlight potential issues such as [overfitting](https://www.ultralytics.com/glossary/overfitting), [underfitting](https://www.ultralytics.com/glossary/underfitting), or inappropriate learning rate settings. By closely monitoring these scalars, you can make informed decisions to optimize the training process, ensuring that the model learns effectively and achieves the desired performance. ### Difference Between Scalars and Time Series @@ -138,15 +138,15 @@ While both Scalars and Time Series in TensorBoard are used for tracking metrics, ### Graphs -The Graphs section of the TensorBoard visualizes the computational graph of the YOLOv8 model, showing how operations and data flow within the model. It's a powerful tool for understanding the model's structure, ensuring that all layers are connected correctly, and for identifying any potential bottlenecks in data flow. Here's an example of what you can expect to see. +The Graphs section of the TensorBoard visualizes the computational graph of the YOLO11 model, showing how operations and data flow within the model. It's a powerful tool for understanding the model's structure, ensuring that all layers are connected correctly, and for identifying any potential bottlenecks in data flow. Here's an example of what you can expect to see. ![image](https://github.com/ultralytics/docs/releases/download/0/tensorboard-yolov8-computational-graph.avif) -Graphs are particularly useful for debugging the model, especially in complex architectures typical in [deep learning](https://www.ultralytics.com/glossary/deep-learning-dl) models like YOLOv8. They help in verifying layer connections and the overall design of the model. +Graphs are particularly useful for debugging the model, especially in complex architectures typical in [deep learning](https://www.ultralytics.com/glossary/deep-learning-dl) models like YOLO11. They help in verifying layer connections and the overall design of the model. ## Summary -This guide aims to help you use TensorBoard with YOLOv8 for visualization and analysis of machine learning model training. It focuses on explaining how key TensorBoard features can provide insights into training metrics and model performance during YOLOv8 training sessions. +This guide aims to help you use TensorBoard with YOLO11 for visualization and analysis of machine learning model training. It focuses on explaining how key TensorBoard features can provide insights into training metrics and model performance during YOLO11 training sessions. For a more detailed exploration of these features and effective utilization strategies, you can refer to TensorFlow's official [TensorBoard documentation](https://www.tensorflow.org/tensorboard/get_started) and their [GitHub repository](https://github.com/tensorflow/tensorboard). @@ -154,29 +154,29 @@ Want to learn more about the various integrations of Ultralytics? Check out the ## FAQ -### What benefits does using TensorBoard with YOLOv8 offer? +### What benefits does using TensorBoard with YOLO11 offer? -Using TensorBoard with YOLOv8 provides several visualization tools essential for efficient model training: +Using TensorBoard with YOLO11 provides several visualization tools essential for efficient model training: - **Real-Time Metrics Tracking:** Track key metrics such as loss, accuracy, precision, and recall live. - **Model Graph Visualization:** Understand and debug the model architecture by visualizing computational graphs. - **Embedding Visualization:** Project embeddings to lower-dimensional spaces for better insight. -These tools enable you to make informed adjustments to enhance your YOLOv8 model's performance. For more details on TensorBoard features, check out the TensorFlow [TensorBoard guide](https://www.tensorflow.org/tensorboard/get_started). +These tools enable you to make informed adjustments to enhance your YOLO11 model's performance. For more details on TensorBoard features, check out the TensorFlow [TensorBoard guide](https://www.tensorflow.org/tensorboard/get_started). -### How can I monitor training metrics using TensorBoard when training a YOLOv8 model? +### How can I monitor training metrics using TensorBoard when training a YOLO11 model? -To monitor training metrics while training a YOLOv8 model with TensorBoard, follow these steps: +To monitor training metrics while training a YOLO11 model with TensorBoard, follow these steps: -1. **Install TensorBoard and YOLOv8:** Run `pip install ultralytics` which includes TensorBoard. -2. **Configure TensorBoard Logging:** During the training process, YOLOv8 logs metrics to a specified log directory. +1. **Install TensorBoard and YOLO11:** Run `pip install ultralytics` which includes TensorBoard. +2. **Configure TensorBoard Logging:** During the training process, YOLO11 logs metrics to a specified log directory. 3. **Start TensorBoard:** Launch TensorBoard using the command `tensorboard --logdir path/to/your/tensorboard/logs`. -The TensorBoard dashboard, accessible via [http://localhost:6006/](http://localhost:6006/), provides real-time insights into various training metrics. For a deeper dive into training configurations, visit our [YOLOv8 Configuration guide](../usage/cfg.md). +The TensorBoard dashboard, accessible via [http://localhost:6006/](http://localhost:6006/), provides real-time insights into various training metrics. For a deeper dive into training configurations, visit our [YOLO11 Configuration guide](../usage/cfg.md). -### What kind of metrics can I visualize with TensorBoard when training YOLOv8 models? +### What kind of metrics can I visualize with TensorBoard when training YOLO11 models? -When training YOLOv8 models, TensorBoard allows you to visualize an array of important metrics including: +When training YOLO11 models, TensorBoard allows you to visualize an array of important metrics including: - **Loss (Training and Validation):** Indicates how well the model is performing during training and validation. - **Accuracy/Precision/[Recall](https://www.ultralytics.com/glossary/recall):** Key performance metrics to evaluate detection accuracy. @@ -185,9 +185,9 @@ When training YOLOv8 models, TensorBoard allows you to visualize an array of imp These visualizations are essential for tracking model performance and making necessary optimizations. For more information on these metrics, refer to our [Performance Metrics guide](../guides/yolo-performance-metrics.md). -### Can I use TensorBoard in a Google Colab environment for training YOLOv8? +### Can I use TensorBoard in a Google Colab environment for training YOLO11? -Yes, you can use TensorBoard in a Google Colab environment to train YOLOv8 models. Here's a quick setup: +Yes, you can use TensorBoard in a Google Colab environment to train YOLO11 models. Here's a quick setup: !!! example "Configure TensorBoard for Google Colab" @@ -198,16 +198,16 @@ Yes, you can use TensorBoard in a Google Colab environment to train YOLOv8 model %tensorboard --logdir path/to/runs ``` - Then, run the YOLOv8 training script: + Then, run the YOLO11 training script: ```python from ultralytics import YOLO # Load a pre-trained model - model = YOLO("yolov8n.pt") + model = YOLO("yolo11n.pt") # Train the model results = model.train(data="coco8.yaml", epochs=100, imgsz=640) ``` -TensorBoard will visualize the training progress within Colab, providing real-time insights into metrics like loss and accuracy. For additional details on configuring YOLOv8 training, see our detailed [YOLOv8 Installation guide](../quickstart.md). +TensorBoard will visualize the training progress within Colab, providing real-time insights into metrics like loss and accuracy. For additional details on configuring YOLO11 training, see our detailed [YOLO11 Installation guide](../quickstart.md). diff --git a/docs/en/integrations/tensorrt.md b/docs/en/integrations/tensorrt.md index 0e40198113..1a8e5a9161 100644 --- a/docs/en/integrations/tensorrt.md +++ b/docs/en/integrations/tensorrt.md @@ -380,7 +380,7 @@ Expand sections below for information on how these models were exported and test See [export mode](../modes/export.md) for details regarding export configuration arguments. - ```py + ```python from ultralytics import YOLO model = YOLO("yolov8n.pt") @@ -401,7 +401,7 @@ Expand sections below for information on how these models were exported and test See [predict mode](../modes/predict.md) for additional information. - ```py + ```python import cv2 from ultralytics import YOLO @@ -421,7 +421,7 @@ Expand sections below for information on how these models were exported and test See [`val` mode](../modes/val.md) to learn more about validation configuration arguments. - ```py + ```python from ultralytics import YOLO model = YOLO("yolov8n.engine") diff --git a/docs/en/integrations/tf-graphdef.md b/docs/en/integrations/tf-graphdef.md index 15cbd48426..fd6d86a32a 100644 --- a/docs/en/integrations/tf-graphdef.md +++ b/docs/en/integrations/tf-graphdef.md @@ -1,14 +1,14 @@ --- comments: true -description: Learn how to export YOLOv8 models to the TF GraphDef format for seamless deployment on various platforms, including mobile and web. -keywords: YOLOv8, export, TensorFlow, GraphDef, model deployment, TensorFlow Serving, TensorFlow Lite, TensorFlow.js, machine learning, AI, computer vision +description: Learn how to export YOLO11 models to the TF GraphDef format for seamless deployment on various platforms, including mobile and web. +keywords: YOLO11, export, TensorFlow, GraphDef, model deployment, TensorFlow Serving, TensorFlow Lite, TensorFlow.js, machine learning, AI, computer vision --- -# How to Export to TF GraphDef from YOLOv8 for Deployment +# How to Export to TF GraphDef from YOLO11 for Deployment -When you are deploying cutting-edge [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models, like YOLOv8, in different environments, you might run into compatibility issues. Google's [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) GraphDef, or TF GraphDef, offers a solution by providing a serialized, platform-independent representation of your model. Using the TF GraphDef model format, you can deploy your YOLOv8 model in environments where the complete TensorFlow ecosystem may not be available, such as mobile devices or specialized hardware. +When you are deploying cutting-edge [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models, like YOLO11, in different environments, you might run into compatibility issues. Google's [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) GraphDef, or TF GraphDef, offers a solution by providing a serialized, platform-independent representation of your model. Using the TF GraphDef model format, you can deploy your YOLO11 model in environments where the complete TensorFlow ecosystem may not be available, such as mobile devices or specialized hardware. -In this guide, we'll walk you step by step through how to export your [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models to the TF GraphDef model format. By converting your model, you can streamline deployment and use YOLOv8's computer vision capabilities in a broader range of applications and platforms. +In this guide, we'll walk you step by step through how to export your [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models to the TF GraphDef model format. By converting your model, you can streamline deployment and use YOLO11's computer vision capabilities in a broader range of applications and platforms.
@@ -16,11 +16,11 @@ In this guide, we'll walk you step by step through how to export your [Ultralyti ## Why Should You Export to TF GraphDef? -TF GraphDef is a powerful component of the TensorFlow ecosystem that was developed by Google. It can be used to optimize and deploy models like YOLOv8. Exporting to TF GraphDef lets us move models from research to real-world applications. It allows models to run in environments without the full TensorFlow framework. +TF GraphDef is a powerful component of the TensorFlow ecosystem that was developed by Google. It can be used to optimize and deploy models like YOLO11. Exporting to TF GraphDef lets us move models from research to real-world applications. It allows models to run in environments without the full TensorFlow framework. The GraphDef format represents the model as a serialized computation graph. This enables various optimization techniques like constant folding, quantization, and graph transformations. These optimizations ensure efficient execution, reduced memory usage, and faster inference speeds. -GraphDef models can use hardware accelerators such as GPUs, TPUs, and AI chips, unlocking significant performance gains for the YOLOv8 inference pipeline. The TF GraphDef format creates a self-contained package with the model and its dependencies, simplifying deployment and integration into diverse systems. +GraphDef models can use hardware accelerators such as GPUs, TPUs, and AI chips, unlocking significant performance gains for the YOLO11 inference pipeline. The TF GraphDef format creates a self-contained package with the model and its dependencies, simplifying deployment and integration into diverse systems. ## Key Features of TF GraphDef Models @@ -38,7 +38,7 @@ Here's a look at its key characteristics: ## Deployment Options with TF GraphDef -Before we dive into the process of exporting YOLOv8 models to TF GraphDef, let's take a look at some typical deployment situations where this format is used. +Before we dive into the process of exporting YOLO11 models to TF GraphDef, let's take a look at some typical deployment situations where this format is used. Here's how you can deploy with TF GraphDef efficiently across various platforms. @@ -46,13 +46,13 @@ Here's how you can deploy with TF GraphDef efficiently across various platforms. - **Mobile and Embedded Devices:** With tools like TensorFlow Lite, you can convert TF GraphDef models into formats optimized for smartphones, tablets, and various embedded devices. Your models can then be used for on-device inference, where execution is done locally, often providing performance gains and offline capabilities. -- **Web Browsers:** TensorFlow.js enables the deployment of TF GraphDef models directly within web browsers. It paves the way for real-time object detection applications running on the client side, using the capabilities of YOLOv8 through JavaScript. +- **Web Browsers:** TensorFlow.js enables the deployment of TF GraphDef models directly within web browsers. It paves the way for real-time object detection applications running on the client side, using the capabilities of YOLO11 through JavaScript. - **Specialized Hardware:** TF GraphDef's platform-agnostic nature allows it to target custom hardware, such as accelerators and TPUs (Tensor Processing Units). These devices can provide performance advantages for computationally intensive models. -## Exporting YOLOv8 Models to TF GraphDef +## Exporting YOLO11 Models to TF GraphDef -You can convert your YOLOv8 object detection model to the TF GraphDef format, which is compatible with various systems, to improve its performance across platforms. +You can convert your YOLO11 object detection model to the TF GraphDef format, which is compatible with various systems, to improve its performance across platforms. ### Installation @@ -63,15 +63,15 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLOv8 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). +Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLO11 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). !!! example "Usage" @@ -80,14 +80,14 @@ Before diving into the usage instructions, it's important to note that while all ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF GraphDef format - model.export(format="pb") # creates 'yolov8n.pb' + model.export(format="pb") # creates 'yolo11n.pb' # Load the exported TF GraphDef model - tf_graphdef_model = YOLO("yolov8n.pb") + tf_graphdef_model = YOLO("yolo11n.pb") # Run inference results = tf_graphdef_model("https://ultralytics.com/images/bus.jpg") @@ -96,18 +96,18 @@ Before diving into the usage instructions, it's important to note that while all === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TF GraphDef format - yolo export model=yolov8n.pt format=pb # creates 'yolov8n.pb' + # Export a YOLO11n PyTorch model to TF GraphDef format + yolo export model=yolo11n.pt format=pb # creates 'yolo11n.pb' # Run inference with the exported model - yolo predict model='yolov8n.pb' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='yolo11n.pb' source='https://ultralytics.com/images/bus.jpg' ``` For more details about supported export options, visit the [Ultralytics documentation page on deployment options](../guides/model-deployment-options.md). -## Deploying Exported YOLOv8 TF GraphDef Models +## Deploying Exported YOLO11 TF GraphDef Models -Once you've exported your YOLOv8 model to the TF GraphDef format, the next step is deployment. The primary and recommended first step for running a TF GraphDef model is to use the YOLO("model.pb") method, as previously shown in the usage code snippet. +Once you've exported your YOLO11 model to the TF GraphDef format, the next step is deployment. The primary and recommended first step for running a TF GraphDef model is to use the YOLO("model.pb") method, as previously shown in the usage code snippet. However, for more information on deploying your TF GraphDef models, take a look at the following resources: @@ -119,17 +119,17 @@ However, for more information on deploying your TF GraphDef models, take a look ## Summary -In this guide, we explored how to export Ultralytics YOLOv8 models to the TF GraphDef format. By doing this, you can flexibly deploy your optimized YOLOv8 models in different environments. +In this guide, we explored how to export Ultralytics YOLO11 models to the TF GraphDef format. By doing this, you can flexibly deploy your optimized YOLO11 models in different environments. For further details on usage, visit the [TF GraphDef official documentation](https://www.tensorflow.org/api_docs/python/tf/Graph). -For more information on integrating Ultralytics YOLOv8 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It has great resources and insights to help you make the most of YOLOv8 in your projects. +For more information on integrating Ultralytics YOLO11 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It has great resources and insights to help you make the most of YOLO11 in your projects. ## FAQ -### How do I export a YOLOv8 model to TF GraphDef format? +### How do I export a YOLO11 model to TF GraphDef format? -Ultralytics YOLOv8 models can be exported to TensorFlow GraphDef (TF GraphDef) format seamlessly. This format provides a serialized, platform-independent representation of the model, ideal for deploying in varied environments like mobile and web. To export a YOLOv8 model to TF GraphDef, follow these steps: +Ultralytics YOLO11 models can be exported to TensorFlow GraphDef (TF GraphDef) format seamlessly. This format provides a serialized, platform-independent representation of the model, ideal for deploying in varied environments like mobile and web. To export a YOLO11 model to TF GraphDef, follow these steps: !!! example "Usage" @@ -138,14 +138,14 @@ Ultralytics YOLOv8 models can be exported to TensorFlow GraphDef (TF GraphDef) f ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF GraphDef format - model.export(format="pb") # creates 'yolov8n.pb' + model.export(format="pb") # creates 'yolo11n.pb' # Load the exported TF GraphDef model - tf_graphdef_model = YOLO("yolov8n.pb") + tf_graphdef_model = YOLO("yolo11n.pb") # Run inference results = tf_graphdef_model("https://ultralytics.com/images/bus.jpg") @@ -154,18 +154,18 @@ Ultralytics YOLOv8 models can be exported to TensorFlow GraphDef (TF GraphDef) f === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TF GraphDef format - yolo export model="yolov8n.pt" format="pb" # creates 'yolov8n.pb' + # Export a YOLO11n PyTorch model to TF GraphDef format + yolo export model="yolo11n.pt" format="pb" # creates 'yolo11n.pb' # Run inference with the exported model - yolo predict model="yolov8n.pb" source="https://ultralytics.com/images/bus.jpg" + yolo predict model="yolo11n.pb" source="https://ultralytics.com/images/bus.jpg" ``` For more information on different export options, visit the [Ultralytics documentation on model export](../modes/export.md). -### What are the benefits of using TF GraphDef for YOLOv8 model deployment? +### What are the benefits of using TF GraphDef for YOLO11 model deployment? -Exporting YOLOv8 models to the TF GraphDef format offers multiple advantages, including: +Exporting YOLO11 models to the TF GraphDef format offers multiple advantages, including: 1. **Platform Independence**: TF GraphDef provides a platform-independent format, allowing models to be deployed across various environments including mobile and web browsers. 2. **Optimizations**: The format enables several optimizations, such as constant folding, quantization, and graph transformations, which enhance execution efficiency and reduce memory usage. @@ -173,19 +173,19 @@ Exporting YOLOv8 models to the TF GraphDef format offers multiple advantages, in Read more about the benefits in the [TF GraphDef section](#why-should-you-export-to-tf-graphdef) of our documentation. -### Why should I use Ultralytics YOLOv8 over other [object detection](https://www.ultralytics.com/glossary/object-detection) models? +### Why should I use Ultralytics YOLO11 over other [object detection](https://www.ultralytics.com/glossary/object-detection) models? -Ultralytics YOLOv8 offers numerous advantages compared to other models like YOLOv5 and YOLOv7. Some key benefits include: +Ultralytics YOLO11 offers numerous advantages compared to other models like YOLOv5 and YOLOv7. Some key benefits include: -1. **State-of-the-Art Performance**: YOLOv8 provides exceptional speed and [accuracy](https://www.ultralytics.com/glossary/accuracy) for real-time object detection, segmentation, and classification. +1. **State-of-the-Art Performance**: YOLO11 provides exceptional speed and [accuracy](https://www.ultralytics.com/glossary/accuracy) for real-time object detection, segmentation, and classification. 2. **Ease of Use**: Features a user-friendly API for model training, validation, prediction, and export, making it accessible for both beginners and experts. 3. **Broad Compatibility**: Supports multiple export formats including ONNX, TensorRT, CoreML, and TensorFlow, for versatile deployment options. -Explore further details in our [introduction to YOLOv8](https://docs.ultralytics.com/models/yolov8/). +Explore further details in our [introduction to YOLO11](https://docs.ultralytics.com/models/yolov8/). -### How can I deploy a YOLOv8 model on specialized hardware using TF GraphDef? +### How can I deploy a YOLO11 model on specialized hardware using TF GraphDef? -Once a YOLOv8 model is exported to TF GraphDef format, you can deploy it across various specialized hardware platforms. Typical deployment scenarios include: +Once a YOLO11 model is exported to TF GraphDef format, you can deploy it across various specialized hardware platforms. Typical deployment scenarios include: - **TensorFlow Serving**: Use TensorFlow Serving for scalable model deployment in production environments. It supports model management and efficient serving. - **Mobile Devices**: Convert TF GraphDef models to TensorFlow Lite, optimized for mobile and embedded devices, enabling on-device inference. @@ -194,11 +194,11 @@ Once a YOLOv8 model is exported to TF GraphDef format, you can deploy it across Check the [deployment options](#deployment-options-with-tf-graphdef) section for detailed information. -### Where can I find solutions for common issues while exporting YOLOv8 models? +### Where can I find solutions for common issues while exporting YOLO11 models? -For troubleshooting common issues with exporting YOLOv8 models, Ultralytics provides comprehensive guides and resources. If you encounter problems during installation or model export, refer to: +For troubleshooting common issues with exporting YOLO11 models, Ultralytics provides comprehensive guides and resources. If you encounter problems during installation or model export, refer to: - **[Common Issues Guide](../guides/yolo-common-issues.md)**: Offers solutions to frequently faced problems. - **[Installation Guide](../quickstart.md)**: Step-by-step instructions for setting up the required packages. -These resources should help you resolve most issues related to YOLOv8 model export and deployment. +These resources should help you resolve most issues related to YOLO11 model export and deployment. diff --git a/docs/en/integrations/tf-savedmodel.md b/docs/en/integrations/tf-savedmodel.md index 9f04dc7893..288802b641 100644 --- a/docs/en/integrations/tf-savedmodel.md +++ b/docs/en/integrations/tf-savedmodel.md @@ -1,14 +1,14 @@ --- comments: true -description: Learn how to export Ultralytics YOLOv8 models to TensorFlow SavedModel format for easy deployment across various platforms and environments. -keywords: YOLOv8, TF SavedModel, Ultralytics, TensorFlow, model export, model deployment, machine learning, AI +description: Learn how to export Ultralytics YOLO11 models to TensorFlow SavedModel format for easy deployment across various platforms and environments. +keywords: YOLO11, TF SavedModel, Ultralytics, TensorFlow, model export, model deployment, machine learning, AI --- -# Understand How to Export to TF SavedModel Format From YOLOv8 +# Understand How to Export to TF SavedModel Format From YOLO11 Deploying [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) models can be challenging. However, using an efficient and flexible model format can make your job easier. TF SavedModel is an open-source machine-learning framework used by TensorFlow to load machine-learning models in a consistent way. It is like a suitcase for TensorFlow models, making them easy to carry and use on different devices and systems. -Learning how to export to TF SavedModel from [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models can help you deploy models easily across different platforms and environments. In this guide, we'll walk through how to convert your models to the TF SavedModel format, simplifying the process of running inferences with your models on different devices. +Learning how to export to TF SavedModel from [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models can help you deploy models easily across different platforms and environments. In this guide, we'll walk through how to convert your models to the TF SavedModel format, simplifying the process of running inferences with your models on different devices. ## Why Should You Export to TF SavedModel? @@ -32,7 +32,7 @@ Here are the key features that make TF SavedModel a great option for AI develope ## Deployment Options with TF SavedModel -Before we dive into the process of exporting YOLOv8 models to the TF SavedModel format, let's explore some typical deployment scenarios where this format is used. +Before we dive into the process of exporting YOLO11 models to the TF SavedModel format, let's explore some typical deployment scenarios where this format is used. TF SavedModel provides a range of options to deploy your machine learning models: @@ -44,9 +44,9 @@ TF SavedModel provides a range of options to deploy your machine learning models - **TensorFlow Runtime:** TensorFlow Runtime (`tfrt`) is a high-performance runtime for executing [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) graphs. It provides lower-level APIs for loading and running TF SavedModels in C++ environments. TensorFlow Runtime offers better performance compared to the standard TensorFlow runtime. It is suitable for deployment scenarios that require low-latency inference and tight integration with existing C++ codebases. -## Exporting YOLOv8 Models to TF SavedModel +## Exporting YOLO11 Models to TF SavedModel -By exporting YOLOv8 models to the TF SavedModel format, you enhance their adaptability and ease of deployment across various platforms. +By exporting YOLO11 models to the TF SavedModel format, you enhance their adaptability and ease of deployment across various platforms. ### Installation @@ -57,15 +57,15 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLOv8 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). +Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLO11 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). !!! example "Usage" @@ -74,14 +74,14 @@ Before diving into the usage instructions, it's important to note that while all ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF SavedModel format - model.export(format="saved_model") # creates '/yolov8n_saved_model' + model.export(format="saved_model") # creates '/yolo11n_saved_model' # Load the exported TF SavedModel model - tf_savedmodel_model = YOLO("./yolov8n_saved_model") + tf_savedmodel_model = YOLO("./yolo11n_saved_model") # Run inference results = tf_savedmodel_model("https://ultralytics.com/images/bus.jpg") @@ -90,18 +90,18 @@ Before diving into the usage instructions, it's important to note that while all === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TF SavedModel format - yolo export model=yolov8n.pt format=saved_model # creates '/yolov8n_saved_model' + # Export a YOLO11n PyTorch model to TF SavedModel format + yolo export model=yolo11n.pt format=saved_model # creates '/yolo11n_saved_model' # Run inference with the exported model - yolo predict model='./yolov8n_saved_model' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='./yolo11n_saved_model' source='https://ultralytics.com/images/bus.jpg' ``` For more details about supported export options, visit the [Ultralytics documentation page on deployment options](../guides/model-deployment-options.md). -## Deploying Exported YOLOv8 TF SavedModel Models +## Deploying Exported YOLO11 TF SavedModel Models -Now that you have exported your YOLOv8 model to the TF SavedModel format, the next step is to deploy it. The primary and recommended first step for running a TF GraphDef model is to use the YOLO("./yolov8n_saved_model") method, as previously shown in the usage code snippet. +Now that you have exported your YOLO11 model to the TF SavedModel format, the next step is to deploy it. The primary and recommended first step for running a TF GraphDef model is to use the YOLO("./yolo11n_saved_model") method, as previously shown in the usage code snippet. However, for in-depth instructions on deploying your TF SavedModel models, take a look at the following resources: @@ -113,11 +113,11 @@ However, for in-depth instructions on deploying your TF SavedModel models, take ## Summary -In this guide, we explored how to export Ultralytics YOLOv8 models to the TF SavedModel format. By exporting to TF SavedModel, you gain the flexibility to optimize, deploy, and scale your YOLOv8 models on a wide range of platforms. +In this guide, we explored how to export Ultralytics YOLO11 models to the TF SavedModel format. By exporting to TF SavedModel, you gain the flexibility to optimize, deploy, and scale your YOLO11 models on a wide range of platforms. For further details on usage, visit the [TF SavedModel official documentation](https://www.tensorflow.org/guide/saved_model). -For more information on integrating Ultralytics YOLOv8 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It's packed with great resources to help you make the most of YOLOv8 in your projects. +For more information on integrating Ultralytics YOLO11 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It's packed with great resources to help you make the most of YOLO11 in your projects. ## FAQ @@ -125,32 +125,32 @@ For more information on integrating Ultralytics YOLOv8 with other platforms and Exporting an Ultralytics YOLO model to the TensorFlow SavedModel format is straightforward. You can use either Python or CLI to achieve this: -!!! example "Exporting YOLOv8 to TF SavedModel" +!!! example "Exporting YOLO11 to TF SavedModel" === "Python" ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF SavedModel format - model.export(format="saved_model") # creates '/yolov8n_saved_model' + model.export(format="saved_model") # creates '/yolo11n_saved_model' # Load the exported TF SavedModel for inference - tf_savedmodel_model = YOLO("./yolov8n_saved_model") + tf_savedmodel_model = YOLO("./yolo11n_saved_model") results = tf_savedmodel_model("https://ultralytics.com/images/bus.jpg") ``` === "CLI" ```bash - # Export the YOLOv8 model to TF SavedModel format - yolo export model=yolov8n.pt format=saved_model # creates '/yolov8n_saved_model' + # Export the YOLO11 model to TF SavedModel format + yolo export model=yolo11n.pt format=saved_model # creates '/yolo11n_saved_model' # Run inference with the exported model - yolo predict model='./yolov8n_saved_model' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='./yolo11n_saved_model' source='https://ultralytics.com/images/bus.jpg' ``` Refer to the [Ultralytics Export documentation](../modes/export.md) for more details. @@ -176,9 +176,9 @@ TF SavedModel can be deployed in various environments, including: For detailed deployment options, visit the official guides on [deploying TensorFlow models](https://www.tensorflow.org/tfx/guide/serving). -### How can I install the necessary packages to export YOLOv8 models? +### How can I install the necessary packages to export YOLO11 models? -To export YOLOv8 models, you need to install the `ultralytics` package. Run the following command in your terminal: +To export YOLO11 models, you need to install the `ultralytics` package. Run the following command in your terminal: ```bash pip install ultralytics diff --git a/docs/en/integrations/tfjs.md b/docs/en/integrations/tfjs.md index ea2d613c62..a8168215b6 100644 --- a/docs/en/integrations/tfjs.md +++ b/docs/en/integrations/tfjs.md @@ -1,14 +1,14 @@ --- comments: true -description: Convert your Ultralytics YOLOv8 models to TensorFlow.js for high-speed, local object detection. Learn how to optimize ML models for browser and Node.js apps. -keywords: YOLOv8, TensorFlow.js, TF.js, model export, machine learning, object detection, browser ML, Node.js, Ultralytics, YOLO, export models +description: Convert your Ultralytics YOLO11 models to TensorFlow.js for high-speed, local object detection. Learn how to optimize ML models for browser and Node.js apps. +keywords: YOLO11, TensorFlow.js, TF.js, model export, machine learning, object detection, browser ML, Node.js, Ultralytics, YOLO, export models --- -# Export to TF.js Model Format From a YOLOv8 Model Format +# Export to TF.js Model Format From a YOLO11 Model Format Deploying [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) models directly in the browser or on Node.js can be tricky. You'll need to make sure your model format is optimized for faster performance so that the model can be used to run interactive applications locally on the user's device. The TensorFlow.js, or TF.js, model format is designed to use minimal power while delivering fast performance. -The 'export to TF.js model format' feature allows you to optimize your [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models for high-speed and locally-run [object detection](https://www.ultralytics.com/glossary/object-detection) inference. In this guide, we'll walk you through converting your models to the TF.js format, making it easier for your models to perform well on various local browsers and Node.js applications. +The 'export to TF.js model format' feature allows you to optimize your [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models for high-speed and locally-run [object detection](https://www.ultralytics.com/glossary/object-detection) inference. In this guide, we'll walk you through converting your models to the TF.js format, making it easier for your models to perform well on various local browsers and Node.js applications. ## Why Should You Export to TF.js? @@ -32,7 +32,7 @@ Here are the key features that make TF.js a powerful tool for developers: ## Deployment Options with TensorFlow.js -Before we dive into the process of exporting YOLOv8 models to the TF.js format, let's explore some typical deployment scenarios where this format is used. +Before we dive into the process of exporting YOLO11 models to the TF.js format, let's explore some typical deployment scenarios where this format is used. TF.js provides a range of options to deploy your machine learning models: @@ -42,9 +42,9 @@ TF.js provides a range of options to deploy your machine learning models: - **Chrome Extensions:** An interesting deployment scenario is the creation of Chrome extensions with TensorFlow.js. For instance, you can develop an extension that allows users to right-click on an image within any webpage to classify it using a pre-trained ML model. TensorFlow.js can be integrated into everyday web browsing experiences to provide immediate insights or augmentations based on machine learning. -## Exporting YOLOv8 Models to TensorFlow.js +## Exporting YOLO11 Models to TensorFlow.js -You can expand model compatibility and deployment flexibility by converting YOLOv8 models to TF.js. +You can expand model compatibility and deployment flexibility by converting YOLO11 models to TF.js. ### Installation @@ -55,15 +55,15 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLOv8 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). +Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLO11 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). !!! example "Usage" @@ -72,14 +72,14 @@ Before diving into the usage instructions, it's important to note that while all ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF.js format - model.export(format="tfjs") # creates '/yolov8n_web_model' + model.export(format="tfjs") # creates '/yolo11n_web_model' # Load the exported TF.js model - tfjs_model = YOLO("./yolov8n_web_model") + tfjs_model = YOLO("./yolo11n_web_model") # Run inference results = tfjs_model("https://ultralytics.com/images/bus.jpg") @@ -88,18 +88,18 @@ Before diving into the usage instructions, it's important to note that while all === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TF.js format - yolo export model=yolov8n.pt format=tfjs # creates '/yolov8n_web_model' + # Export a YOLO11n PyTorch model to TF.js format + yolo export model=yolo11n.pt format=tfjs # creates '/yolo11n_web_model' # Run inference with the exported model - yolo predict model='./yolov8n_web_model' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='./yolo11n_web_model' source='https://ultralytics.com/images/bus.jpg' ``` For more details about supported export options, visit the [Ultralytics documentation page on deployment options](../guides/model-deployment-options.md). -## Deploying Exported YOLOv8 TensorFlow.js Models +## Deploying Exported YOLO11 TensorFlow.js Models -Now that you have exported your YOLOv8 model to the TF.js format, the next step is to deploy it. The primary and recommended first step for running a TF.js is to use the YOLO("./yolov8n_web_model") method, as previously shown in the usage code snippet. +Now that you have exported your YOLO11 model to the TF.js format, the next step is to deploy it. The primary and recommended first step for running a TF.js is to use the `YOLO("./yolo11n_web_model")` method, as previously shown in the usage code snippet. However, for in-depth instructions on deploying your TF.js models, take a look at the following resources: @@ -111,17 +111,17 @@ However, for in-depth instructions on deploying your TF.js models, take a look a ## Summary -In this guide, we learned how to export Ultralytics YOLOv8 models to the TensorFlow.js format. By exporting to TF.js, you gain the flexibility to optimize, deploy, and scale your YOLOv8 models on a wide range of platforms. +In this guide, we learned how to export Ultralytics YOLO11 models to the TensorFlow.js format. By exporting to TF.js, you gain the flexibility to optimize, deploy, and scale your YOLO11 models on a wide range of platforms. For further details on usage, visit the [TensorFlow.js official documentation](https://www.tensorflow.org/js/guide). -For more information on integrating Ultralytics YOLOv8 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It's packed with great resources to help you make the most of YOLOv8 in your projects. +For more information on integrating Ultralytics YOLO11 with other platforms and frameworks, don't forget to check out our [integration guide page](index.md). It's packed with great resources to help you make the most of YOLO11 in your projects. ## FAQ -### How do I export Ultralytics YOLOv8 models to TensorFlow.js format? +### How do I export Ultralytics YOLO11 models to TensorFlow.js format? -Exporting Ultralytics YOLOv8 models to TensorFlow.js (TF.js) format is straightforward. You can follow these steps: +Exporting Ultralytics YOLO11 models to TensorFlow.js (TF.js) format is straightforward. You can follow these steps: !!! example "Usage" @@ -130,14 +130,14 @@ Exporting Ultralytics YOLOv8 models to TensorFlow.js (TF.js) format is straightf ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TF.js format - model.export(format="tfjs") # creates '/yolov8n_web_model' + model.export(format="tfjs") # creates '/yolo11n_web_model' # Load the exported TF.js model - tfjs_model = YOLO("./yolov8n_web_model") + tfjs_model = YOLO("./yolo11n_web_model") # Run inference results = tfjs_model("https://ultralytics.com/images/bus.jpg") @@ -146,18 +146,18 @@ Exporting Ultralytics YOLOv8 models to TensorFlow.js (TF.js) format is straightf === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TF.js format - yolo export model=yolov8n.pt format=tfjs # creates '/yolov8n_web_model' + # Export a YOLO11n PyTorch model to TF.js format + yolo export model=yolo11n.pt format=tfjs # creates '/yolo11n_web_model' # Run inference with the exported model - yolo predict model='./yolov8n_web_model' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='./yolo11n_web_model' source='https://ultralytics.com/images/bus.jpg' ``` For more details about supported export options, visit the [Ultralytics documentation page on deployment options](../guides/model-deployment-options.md). -### Why should I export my YOLOv8 models to TensorFlow.js? +### Why should I export my YOLO11 models to TensorFlow.js? -Exporting YOLOv8 models to TensorFlow.js offers several advantages, including: +Exporting YOLO11 models to TensorFlow.js offers several advantages, including: 1. **Local Execution:** Models can run directly in the browser or Node.js, reducing latency and enhancing user experience. 2. **Cross-Platform Support:** TF.js supports multiple environments, allowing flexibility in deployment. @@ -177,7 +177,7 @@ TensorFlow.js is specifically designed for efficient execution of ML models in b Interested in learning more about TF.js? Check out the [official TensorFlow.js guide](https://www.tensorflow.org/js/guide). -### What are the key features of TensorFlow.js for deploying YOLOv8 models? +### What are the key features of TensorFlow.js for deploying YOLO11 models? Key features of TensorFlow.js include: @@ -185,10 +185,10 @@ Key features of TensorFlow.js include: - **Multiple Backends:** Supports CPU, WebGL for GPU acceleration, WebAssembly (WASM), and WebGPU for advanced operations. - **Offline Capabilities:** Models can run directly in the browser without internet connectivity, making it ideal for developing responsive web applications. -For deployment scenarios and more in-depth information, see our section on [Deployment Options with TensorFlow.js](#deploying-exported-yolov8-tensorflowjs-models). +For deployment scenarios and more in-depth information, see our section on [Deployment Options with TensorFlow.js](#deploying-exported-yolo11-tensorflowjs-models). -### Can I deploy a YOLOv8 model on server-side Node.js applications using TensorFlow.js? +### Can I deploy a YOLO11 model on server-side Node.js applications using TensorFlow.js? -Yes, TensorFlow.js allows the deployment of YOLOv8 models on Node.js environments. This enables server-side machine learning applications that benefit from the processing power of a server and access to server-side data. Typical use cases include real-time data processing and machine learning pipelines on backend servers. +Yes, TensorFlow.js allows the deployment of YOLO11 models on Node.js environments. This enables server-side machine learning applications that benefit from the processing power of a server and access to server-side data. Typical use cases include real-time data processing and machine learning pipelines on backend servers. To get started with Node.js deployment, refer to the [Run TensorFlow.js in Node.js](https://www.tensorflow.org/js/guide/nodejs) guide from TensorFlow. diff --git a/docs/en/integrations/tflite.md b/docs/en/integrations/tflite.md index 028675eabb..9f0ebad156 100644 --- a/docs/en/integrations/tflite.md +++ b/docs/en/integrations/tflite.md @@ -1,10 +1,10 @@ --- comments: true -description: Learn how to convert YOLOv8 models to TFLite for edge device deployment. Optimize performance and ensure seamless execution on various platforms. -keywords: YOLOv8, TFLite, model export, TensorFlow Lite, edge devices, deployment, Ultralytics, machine learning, on-device inference, model optimization +description: Learn how to convert YOLO11 models to TFLite for edge device deployment. Optimize performance and ensure seamless execution on various platforms. +keywords: YOLO11, TFLite, model export, TensorFlow Lite, edge devices, deployment, Ultralytics, machine learning, on-device inference, model optimization --- -# A Guide on YOLOv8 Model Export to TFLite for Deployment +# A Guide on YOLO11 Model Export to TFLite for Deployment
@@ -12,7 +12,7 @@ keywords: YOLOv8, TFLite, model export, TensorFlow Lite, edge devices, deploymen Deploying [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models on edge devices or embedded devices requires a format that can ensure seamless performance. -The TensorFlow Lite or TFLite export format allows you to optimize your [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models for tasks like [object detection](https://www.ultralytics.com/glossary/object-detection) and [image classification](https://www.ultralytics.com/glossary/image-classification) in edge device-based applications. In this guide, we'll walk through the steps for converting your models to the TFLite format, making it easier for your models to perform well on various edge devices. +The TensorFlow Lite or TFLite export format allows you to optimize your [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models for tasks like [object detection](https://www.ultralytics.com/glossary/object-detection) and [image classification](https://www.ultralytics.com/glossary/image-classification) in edge device-based applications. In this guide, we'll walk through the steps for converting your models to the TFLite format, making it easier for your models to perform well on various edge devices. ## Why should you export to TFLite? @@ -34,7 +34,7 @@ TFLite models offer a wide range of key features that enable on-device machine l ## Deployment Options in TFLite -Before we look at the code for exporting YOLOv8 models to the TFLite format, let's understand how TFLite models are normally used. +Before we look at the code for exporting YOLO11 models to the TFLite format, let's understand how TFLite models are normally used. TFLite offers various on-device deployment options for machine learning models, including: @@ -48,7 +48,7 @@ TFLite offers various on-device deployment options for machine learning models, - **Deploying with Microcontrollers**: TFLite models can also be deployed on microcontrollers and other devices with only a few kilobytes of memory. The core runtime just fits in 16 KB on an Arm Cortex M3 and can run many basic models. It doesn't require operating system support, any standard C or C++ libraries, or dynamic memory allocation. -## Export to TFLite: Converting Your YOLOv8 Model +## Export to TFLite: Converting Your YOLO11 Model You can improve on-device model execution efficiency and optimize performance by converting them to TFLite format. @@ -61,15 +61,15 @@ To install the required packages, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLOv8 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). +Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLO11 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). !!! example "Usage" @@ -78,14 +78,14 @@ Before diving into the usage instructions, it's important to note that while all ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TFLite format - model.export(format="tflite") # creates 'yolov8n_float32.tflite' + model.export(format="tflite") # creates 'yolo11n_float32.tflite' # Load the exported TFLite model - tflite_model = YOLO("yolov8n_float32.tflite") + tflite_model = YOLO("yolo11n_float32.tflite") # Run inference results = tflite_model("https://ultralytics.com/images/bus.jpg") @@ -94,18 +94,18 @@ Before diving into the usage instructions, it's important to note that while all === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TFLite format - yolo export model=yolov8n.pt format=tflite # creates 'yolov8n_float32.tflite' + # Export a YOLO11n PyTorch model to TFLite format + yolo export model=yolo11n.pt format=tflite # creates 'yolo11n_float32.tflite' # Run inference with the exported model - yolo predict model='yolov8n_float32.tflite' source='https://ultralytics.com/images/bus.jpg' + yolo predict model='yolo11n_float32.tflite' source='https://ultralytics.com/images/bus.jpg' ``` For more details about the export process, visit the [Ultralytics documentation page on exporting](../modes/export.md). -## Deploying Exported YOLOv8 TFLite Models +## Deploying Exported YOLO11 TFLite Models -After successfully exporting your Ultralytics YOLOv8 models to TFLite format, you can now deploy them. The primary and recommended first step for running a TFLite model is to utilize the YOLO("model.tflite") method, as outlined in the previous usage code snippet. However, for in-depth instructions on deploying your TFLite models in various other settings, take a look at the following resources: +After successfully exporting your Ultralytics YOLO11 models to TFLite format, you can now deploy them. The primary and recommended first step for running a TFLite model is to utilize the YOLO("model.tflite") method, as outlined in the previous usage code snippet. However, for in-depth instructions on deploying your TFLite models in various other settings, take a look at the following resources: - **[Android](https://ai.google.dev/edge/litert/android)**: A quick start guide for integrating [TensorFlow](https://www.ultralytics.com/glossary/tensorflow) Lite into Android applications, providing easy-to-follow steps for setting up and running [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) models. @@ -115,17 +115,17 @@ After successfully exporting your Ultralytics YOLOv8 models to TFLite format, yo ## Summary -In this guide, we focused on how to export to TFLite format. By converting your Ultralytics YOLOv8 models to TFLite model format, you can improve the efficiency and speed of YOLOv8 models, making them more effective and suitable for [edge computing](https://www.ultralytics.com/glossary/edge-computing) environments. +In this guide, we focused on how to export to TFLite format. By converting your Ultralytics YOLO11 models to TFLite model format, you can improve the efficiency and speed of YOLO11 models, making them more effective and suitable for [edge computing](https://www.ultralytics.com/glossary/edge-computing) environments. For further details on usage, visit the [TFLite official documentation](https://ai.google.dev/edge/litert). -Also, if you're curious about other Ultralytics YOLOv8 integrations, make sure to check out our [integration guide page](../integrations/index.md). You'll find tons of helpful info and insights waiting for you there. +Also, if you're curious about other Ultralytics YOLO11 integrations, make sure to check out our [integration guide page](../integrations/index.md). You'll find tons of helpful info and insights waiting for you there. ## FAQ -### How do I export a YOLOv8 model to TFLite format? +### How do I export a YOLO11 model to TFLite format? -To export a YOLOv8 model to TFLite format, you can use the Ultralytics library. First, install the required package using: +To export a YOLO11 model to TFLite format, you can use the Ultralytics library. First, install the required package using: ```bash pip install ultralytics @@ -136,24 +136,24 @@ Then, use the following code snippet to export your model: ```python from ultralytics import YOLO -# Load the YOLOv8 model -model = YOLO("yolov8n.pt") +# Load the YOLO11 model +model = YOLO("yolo11n.pt") # Export the model to TFLite format -model.export(format="tflite") # creates 'yolov8n_float32.tflite' +model.export(format="tflite") # creates 'yolo11n_float32.tflite' ``` For CLI users, you can achieve this with: ```bash -yolo export model=yolov8n.pt format=tflite # creates 'yolov8n_float32.tflite' +yolo export model=yolo11n.pt format=tflite # creates 'yolo11n_float32.tflite' ``` For more details, visit the [Ultralytics export guide](../modes/export.md). -### What are the benefits of using TensorFlow Lite for YOLOv8 [model deployment](https://www.ultralytics.com/glossary/model-deployment)? +### What are the benefits of using TensorFlow Lite for YOLO11 [model deployment](https://www.ultralytics.com/glossary/model-deployment)? -TensorFlow Lite (TFLite) is an open-source [deep learning](https://www.ultralytics.com/glossary/deep-learning-dl) framework designed for on-device inference, making it ideal for deploying YOLOv8 models on mobile, embedded, and IoT devices. Key benefits include: +TensorFlow Lite (TFLite) is an open-source [deep learning](https://www.ultralytics.com/glossary/deep-learning-dl) framework designed for on-device inference, making it ideal for deploying YOLO11 models on mobile, embedded, and IoT devices. Key benefits include: - **On-device optimization**: Minimize latency and enhance privacy by processing data locally. - **Platform compatibility**: Supports Android, iOS, embedded Linux, and MCU. @@ -161,33 +161,33 @@ TensorFlow Lite (TFLite) is an open-source [deep learning](https://www.ultralyti To learn more, check out the [TFLite guide](https://ai.google.dev/edge/litert). -### Is it possible to run YOLOv8 TFLite models on Raspberry Pi? +### Is it possible to run YOLO11 TFLite models on Raspberry Pi? -Yes, you can run YOLOv8 TFLite models on Raspberry Pi to improve inference speeds. First, export your model to TFLite format as explained [here](#how-do-i-export-a-yolov8-model-to-tflite-format). Then, use a tool like TensorFlow Lite Interpreter to execute the model on your Raspberry Pi. +Yes, you can run YOLO11 TFLite models on Raspberry Pi to improve inference speeds. First, export your model to TFLite format as explained [here](#how-do-i-export-a-yolo11-model-to-tflite-format). Then, use a tool like TensorFlow Lite Interpreter to execute the model on your Raspberry Pi. For further optimizations, you might consider using [Coral Edge TPU](https://coral.withgoogle.com/). For detailed steps, refer to our [Raspberry Pi deployment guide](../guides/raspberry-pi.md). -### Can I use TFLite models on microcontrollers for YOLOv8 predictions? +### Can I use TFLite models on microcontrollers for YOLO11 predictions? -Yes, TFLite supports deployment on microcontrollers with limited resources. TFLite's core runtime requires only 16 KB of memory on an Arm Cortex M3 and can run basic YOLOv8 models. This makes it suitable for deployment on devices with minimal computational power and memory. +Yes, TFLite supports deployment on microcontrollers with limited resources. TFLite's core runtime requires only 16 KB of memory on an Arm Cortex M3 and can run basic YOLO11 models. This makes it suitable for deployment on devices with minimal computational power and memory. To get started, visit the [TFLite Micro for Microcontrollers guide](https://ai.google.dev/edge/litert/microcontrollers/overview). -### What platforms are compatible with TFLite exported YOLOv8 models? +### What platforms are compatible with TFLite exported YOLO11 models? -TensorFlow Lite provides extensive platform compatibility, allowing you to deploy YOLOv8 models on a wide range of devices, including: +TensorFlow Lite provides extensive platform compatibility, allowing you to deploy YOLO11 models on a wide range of devices, including: - **Android and iOS**: Native support through TFLite Android and iOS libraries. - **Embedded Linux**: Ideal for single-board computers such as Raspberry Pi. - **Microcontrollers**: Suitable for MCUs with constrained resources. -For more information on deployment options, see our detailed [deployment guide](#deploying-exported-yolov8-tflite-models). +For more information on deployment options, see our detailed [deployment guide](#deploying-exported-yolo11-tflite-models). -### How do I troubleshoot common issues during YOLOv8 model export to TFLite? +### How do I troubleshoot common issues during YOLO11 model export to TFLite? -If you encounter errors while exporting YOLOv8 models to TFLite, common solutions include: +If you encounter errors while exporting YOLO11 models to TFLite, common solutions include: - **Check package compatibility**: Ensure you're using compatible versions of Ultralytics and TensorFlow. Refer to our [installation guide](../quickstart.md). -- **Model support**: Verify that the specific YOLOv8 model supports TFLite export by checking [here](../modes/export.md). +- **Model support**: Verify that the specific YOLO11 model supports TFLite export by checking [here](../modes/export.md). For additional troubleshooting tips, visit our [Common Issues guide](../guides/yolo-common-issues.md). diff --git a/docs/en/integrations/torchscript.md b/docs/en/integrations/torchscript.md index 839caff921..1be1516c0b 100644 --- a/docs/en/integrations/torchscript.md +++ b/docs/en/integrations/torchscript.md @@ -1,22 +1,22 @@ --- comments: true -description: Learn how to export Ultralytics YOLOv8 models to TorchScript for flexible, cross-platform deployment. Boost performance and utilize in various environments. -keywords: YOLOv8, TorchScript, model export, Ultralytics, PyTorch, deep learning, AI deployment, cross-platform, performance optimization +description: Learn how to export Ultralytics YOLO11 models to TorchScript for flexible, cross-platform deployment. Boost performance and utilize in various environments. +keywords: YOLO11, TorchScript, model export, Ultralytics, PyTorch, deep learning, AI deployment, cross-platform, performance optimization --- -# YOLOv8 Model Export to TorchScript for Quick Deployment +# YOLO11 Model Export to TorchScript for Quick Deployment Deploying [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) models across different environments, including embedded systems, web browsers, or platforms with limited Python support, requires a flexible and portable solution. TorchScript focuses on portability and the ability to run models in environments where the entire Python framework is unavailable. This makes it ideal for scenarios where you need to deploy your computer vision capabilities across various devices or platforms. -Export to Torchscript to serialize your [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) models for cross-platform compatibility and streamlined deployment. In this guide, we'll show you how to export your YOLOv8 models to the TorchScript format, making it easier for you to use them across a wider range of applications. +Export to Torchscript to serialize your [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) models for cross-platform compatibility and streamlined deployment. In this guide, we'll show you how to export your YOLO11 models to the TorchScript format, making it easier for you to use them across a wider range of applications. ## Why should you export to TorchScript? ![Torchscript Overview](https://github.com/ultralytics/docs/releases/download/0/torchscript-overview.avif) -Developed by the creators of PyTorch, TorchScript is a powerful tool for optimizing and deploying PyTorch models across a variety of platforms. Exporting YOLOv8 models to [TorchScript](https://pytorch.org/docs/stable/jit.html) is crucial for moving from research to real-world applications. TorchScript, part of the PyTorch framework, helps make this transition smoother by allowing PyTorch models to be used in environments that don't support Python. +Developed by the creators of PyTorch, TorchScript is a powerful tool for optimizing and deploying PyTorch models across a variety of platforms. Exporting YOLO11 models to [TorchScript](https://pytorch.org/docs/stable/jit.html) is crucial for moving from research to real-world applications. TorchScript, part of the PyTorch framework, helps make this transition smoother by allowing PyTorch models to be used in environments that don't support Python. -The process involves two techniques: tracing and scripting. Tracing records operations during model execution, while scripting allows for the definition of models using a subset of Python. These techniques ensure that models like YOLOv8 can still work their magic even outside their usual Python environment. +The process involves two techniques: tracing and scripting. Tracing records operations during model execution, while scripting allows for the definition of models using a subset of Python. These techniques ensure that models like YOLO11 can still work their magic even outside their usual Python environment. ![TorchScript Script and Trace](https://github.com/ultralytics/docs/releases/download/0/torchscript-script-and-trace.avif) @@ -42,7 +42,7 @@ Here are the key features that make TorchScript a valuable tool for developers: ## Deployment Options in TorchScript -Before we look at the code for exporting YOLOv8 models to the TorchScript format, let's understand where TorchScript models are normally used. +Before we look at the code for exporting YOLO11 models to the TorchScript format, let's understand where TorchScript models are normally used. TorchScript offers various deployment options for [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) models, such as: @@ -52,9 +52,9 @@ TorchScript offers various deployment options for [machine learning](https://www - **Cloud Deployment**: TorchScript models can be deployed to cloud-based servers using solutions like TorchServe. It provides features like model versioning, batching, and metrics monitoring for scalable deployment in production environments. Cloud deployment with TorchScript can make your models accessible via APIs or other web services. -## Export to TorchScript: Converting Your YOLOv8 Model +## Export to TorchScript: Converting Your YOLO11 Model -Exporting YOLOv8 models to TorchScript makes it easier to use them in different places and helps them run faster and more efficiently. This is great for anyone looking to use deep learning models more effectively in real-world applications. +Exporting YOLO11 models to TorchScript makes it easier to use them in different places and helps them run faster and more efficiently. This is great for anyone looking to use deep learning models more effectively in real-world applications. ### Installation @@ -65,15 +65,15 @@ To install the required package, run: === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` -For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLOv8, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. +For detailed instructions and best practices related to the installation process, check our [Ultralytics Installation guide](../quickstart.md). While installing the required packages for YOLO11, if you encounter any difficulties, consult our [Common Issues guide](../guides/yolo-common-issues.md) for solutions and tips. ### Usage -Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLOv8 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). +Before diving into the usage instructions, it's important to note that while all [Ultralytics YOLO11 models](../models/index.md) are available for exporting, you can ensure that the model you select supports export functionality [here](../modes/export.md). !!! example "Usage" @@ -82,14 +82,14 @@ Before diving into the usage instructions, it's important to note that while all ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TorchScript format - model.export(format="torchscript") # creates 'yolov8n.torchscript' + model.export(format="torchscript") # creates 'yolo11n.torchscript' # Load the exported TorchScript model - torchscript_model = YOLO("yolov8n.torchscript") + torchscript_model = YOLO("yolo11n.torchscript") # Run inference results = torchscript_model("https://ultralytics.com/images/bus.jpg") @@ -98,18 +98,18 @@ Before diving into the usage instructions, it's important to note that while all === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TorchScript format - yolo export model=yolov8n.pt format=torchscript # creates 'yolov8n.torchscript' + # Export a YOLO11n PyTorch model to TorchScript format + yolo export model=yolo11n.pt format=torchscript # creates 'yolo11n.torchscript' # Run inference with the exported model - yolo predict model=yolov8n.torchscript source='https://ultralytics.com/images/bus.jpg' + yolo predict model=yolo11n.torchscript source='https://ultralytics.com/images/bus.jpg' ``` For more details about the export process, visit the [Ultralytics documentation page on exporting](../modes/export.md). -## Deploying Exported YOLOv8 TorchScript Models +## Deploying Exported YOLO11 TorchScript Models -After successfully exporting your Ultralytics YOLOv8 models to TorchScript format, you can now deploy them. The primary and recommended first step for running a TorchScript model is to utilize the YOLO("model.torchscript") method, as outlined in the previous usage code snippet. However, for in-depth instructions on deploying your TorchScript models in various other settings, take a look at the following resources: +After successfully exporting your Ultralytics YOLO11 models to TorchScript format, you can now deploy them. The primary and recommended first step for running a TorchScript model is to utilize the YOLO("model.torchscript") method, as outlined in the previous usage code snippet. However, for in-depth instructions on deploying your TorchScript models in various other settings, take a look at the following resources: - **[Explore Mobile Deployment](https://pytorch.org/mobile/home/)**: The [PyTorch](https://www.ultralytics.com/glossary/pytorch) Mobile Documentation provides comprehensive guidelines for deploying models on mobile devices, ensuring your applications are efficient and responsive. @@ -119,21 +119,21 @@ After successfully exporting your Ultralytics YOLOv8 models to TorchScript forma ## Summary -In this guide, we explored the process of exporting Ultralytics YOLOv8 models to the TorchScript format. By following the provided instructions, you can optimize YOLOv8 models for performance and gain the flexibility to deploy them across various platforms and environments. +In this guide, we explored the process of exporting Ultralytics YOLO11 models to the TorchScript format. By following the provided instructions, you can optimize YOLO11 models for performance and gain the flexibility to deploy them across various platforms and environments. For further details on usage, visit [TorchScript's official documentation](https://pytorch.org/docs/stable/jit.html). -Also, if you'd like to know more about other Ultralytics YOLOv8 integrations, visit our [integration guide page](../integrations/index.md). You'll find plenty of useful resources and insights there. +Also, if you'd like to know more about other Ultralytics YOLO11 integrations, visit our [integration guide page](../integrations/index.md). You'll find plenty of useful resources and insights there. ## FAQ -### What is Ultralytics YOLOv8 model export to TorchScript? +### What is Ultralytics YOLO11 model export to TorchScript? -Exporting an Ultralytics YOLOv8 model to TorchScript allows for flexible, cross-platform deployment. TorchScript, a part of the PyTorch ecosystem, facilitates the serialization of models, which can then be executed in environments that lack Python support. This makes it ideal for deploying models on embedded systems, C++ environments, mobile applications, and even web browsers. Exporting to TorchScript enables efficient performance and wider applicability of your YOLOv8 models across diverse platforms. +Exporting an Ultralytics YOLO11 model to TorchScript allows for flexible, cross-platform deployment. TorchScript, a part of the PyTorch ecosystem, facilitates the serialization of models, which can then be executed in environments that lack Python support. This makes it ideal for deploying models on embedded systems, C++ environments, mobile applications, and even web browsers. Exporting to TorchScript enables efficient performance and wider applicability of your YOLO11 models across diverse platforms. -### How can I export my YOLOv8 model to TorchScript using Ultralytics? +### How can I export my YOLO11 model to TorchScript using Ultralytics? -To export a YOLOv8 model to TorchScript, you can use the following example code: +To export a YOLO11 model to TorchScript, you can use the following example code: !!! example "Usage" @@ -142,14 +142,14 @@ To export a YOLOv8 model to TorchScript, you can use the following example code: ```python from ultralytics import YOLO - # Load the YOLOv8 model - model = YOLO("yolov8n.pt") + # Load the YOLO11 model + model = YOLO("yolo11n.pt") # Export the model to TorchScript format - model.export(format="torchscript") # creates 'yolov8n.torchscript' + model.export(format="torchscript") # creates 'yolo11n.torchscript' # Load the exported TorchScript model - torchscript_model = YOLO("yolov8n.torchscript") + torchscript_model = YOLO("yolo11n.torchscript") # Run inference results = torchscript_model("https://ultralytics.com/images/bus.jpg") @@ -158,18 +158,18 @@ To export a YOLOv8 model to TorchScript, you can use the following example code: === "CLI" ```bash - # Export a YOLOv8n PyTorch model to TorchScript format - yolo export model=yolov8n.pt format=torchscript # creates 'yolov8n.torchscript' + # Export a YOLO11n PyTorch model to TorchScript format + yolo export model=yolo11n.pt format=torchscript # creates 'yolo11n.torchscript' # Run inference with the exported model - yolo predict model=yolov8n.torchscript source='https://ultralytics.com/images/bus.jpg' + yolo predict model=yolo11n.torchscript source='https://ultralytics.com/images/bus.jpg' ``` For more details about the export process, refer to the [Ultralytics documentation on exporting](../modes/export.md). -### Why should I use TorchScript for deploying YOLOv8 models? +### Why should I use TorchScript for deploying YOLO11 models? -Using TorchScript for deploying YOLOv8 models offers several advantages: +Using TorchScript for deploying YOLO11 models offers several advantages: - **Portability**: Exported models can run in environments without the need for Python, such as C++ applications, embedded systems, or mobile devices. - **Optimization**: TorchScript supports static graph execution and Just-In-Time (JIT) compilation, which can optimize model performance. @@ -178,24 +178,24 @@ Using TorchScript for deploying YOLOv8 models offers several advantages: For more insights into deployment, visit the [PyTorch Mobile Documentation](https://pytorch.org/mobile/home/), [TorchServe Documentation](https://pytorch.org/serve/getting_started.html), and [C++ Deployment Guide](https://pytorch.org/tutorials/advanced/cpp_export.html). -### What are the installation steps for exporting YOLOv8 models to TorchScript? +### What are the installation steps for exporting YOLO11 models to TorchScript? -To install the required package for exporting YOLOv8 models, use the following command: +To install the required package for exporting YOLO11 models, use the following command: !!! tip "Installation" === "CLI" ```bash - # Install the required package for YOLOv8 + # Install the required package for YOLO11 pip install ultralytics ``` For detailed instructions, visit the [Ultralytics Installation guide](../quickstart.md). If any issues arise during installation, consult the [Common Issues guide](../guides/yolo-common-issues.md). -### How do I deploy my exported TorchScript YOLOv8 models? +### How do I deploy my exported TorchScript YOLO11 models? -After exporting YOLOv8 models to the TorchScript format, you can deploy them across a variety of platforms: +After exporting YOLO11 models to the TorchScript format, you can deploy them across a variety of platforms: - **C++ API**: Ideal for low-overhead, highly efficient production environments. - **Mobile Deployment**: Use [PyTorch Mobile](https://pytorch.org/mobile/home/) for iOS and Android applications. diff --git a/docs/en/integrations/vscode.md b/docs/en/integrations/vscode.md index b6785d1529..521abde311 100644 --- a/docs/en/integrations/vscode.md +++ b/docs/en/integrations/vscode.md @@ -134,7 +134,7 @@ The `ultra.examples` snippets are to useful for anyone looking to learn how to g ```python from ultralytics import ASSETS, YOLO - model = YOLO("yolov8n.pt", task="detect") + model = YOLO("yolo11n.pt", task="detect") results = model(source=ASSETS / "bus.jpg") for result in results: diff --git a/docs/en/integrations/weights-biases.md b/docs/en/integrations/weights-biases.md index b831f0b72a..9f2cbb2fa0 100644 --- a/docs/en/integrations/weights-biases.md +++ b/docs/en/integrations/weights-biases.md @@ -1,12 +1,12 @@ --- comments: true -description: Learn how to enhance YOLOv8 experiment tracking and visualization with Weights & Biases for better model performance and management. -keywords: YOLOv8, Weights & Biases, model training, experiment tracking, Ultralytics, machine learning, computer vision, model visualization +description: Learn how to enhance YOLO11 experiment tracking and visualization with Weights & Biases for better model performance and management. +keywords: YOLO11, Weights & Biases, model training, experiment tracking, Ultralytics, machine learning, computer vision, model visualization --- -# Enhancing YOLOv8 Experiment Tracking and Visualization with Weights & Biases +# Enhancing YOLO11 Experiment Tracking and Visualization with Weights & Biases -[Object detection](https://www.ultralytics.com/glossary/object-detection) models like [Ultralytics YOLOv8](https://github.com/ultralytics/ultralytics) have become integral to many [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) applications. However, training, evaluating, and deploying these complex models introduces several challenges. Tracking key training metrics, comparing model variants, analyzing model behavior, and detecting issues require substantial instrumentation and experiment management. +[Object detection](https://www.ultralytics.com/glossary/object-detection) models like [Ultralytics YOLO11](https://github.com/ultralytics/ultralytics) have become integral to many [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) applications. However, training, evaluating, and deploying these complex models introduces several challenges. Tracking key training metrics, comparing model variants, analyzing model behavior, and detecting issues require substantial instrumentation and experiment management.
@@ -16,10 +16,10 @@ keywords: YOLOv8, Weights & Biases, model training, experiment tracking, Ultraly
allowfullscreen>
- Watch: How to use Ultralytics YOLOv8 with Weights and Biases
+ Watch: How to use Ultralytics YOLO11 with Weights and Biases
+
+
+
+ Watch: How to Run Inference with MobileSAM using Ultralytics | Step-by-Step Guide 🎉
+
+
+
+
+ Watch: How to Run Inference with Meta's SAM2 using Ultralytics | Step-by-Step Guide 🎉
+
@@ -50,7 +50,7 @@ Once your model is trained and validated, the next logical step is to evaluate i
## Usage Examples
-Run YOLOv8n benchmarks on all supported export formats including ONNX, TensorRT etc. See Arguments section below for a full list of export arguments.
+Run YOLO11n benchmarks on all supported export formats including ONNX, TensorRT etc. See Arguments section below for a full list of export arguments.
!!! example
@@ -60,13 +60,13 @@ Run YOLOv8n benchmarks on all supported export formats including ONNX, TensorRT
from ultralytics.utils.benchmarks import benchmark
# Benchmark on GPU
- benchmark(model="yolov8n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
+ benchmark(model="yolo11n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
```
=== "CLI"
```bash
- yolo benchmark model=yolov8n.pt data='coco8.yaml' imgsz=640 half=False device=0
+ yolo benchmark model=yolo11n.pt data='coco8.yaml' imgsz=640 half=False device=0
```
## Arguments
@@ -75,7 +75,7 @@ Arguments such as `model`, `data`, `imgsz`, `half`, `device`, and `verbose` prov
| Key | Default Value | Description |
| --------- | ------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
-| `model` | `None` | Specifies the path to the model file. Accepts both `.pt` and `.yaml` formats, e.g., `"yolov8n.pt"` for pre-trained models or configuration files. |
+| `model` | `None` | Specifies the path to the model file. Accepts both `.pt` and `.yaml` formats, e.g., `"yolo11n.pt"` for pre-trained models or configuration files. |
| `data` | `None` | Path to a YAML file defining the dataset for benchmarking, typically including paths and settings for [validation data](https://www.ultralytics.com/glossary/validation-data). Example: `"coco8.yaml"`. |
| `imgsz` | `640` | The input image size for the model. Can be a single integer for square images or a tuple `(width, height)` for non-square, e.g., `(640, 480)`. |
| `half` | `False` | Enables FP16 (half-precision) inference, reducing memory usage and possibly increasing speed on compatible hardware. Use `half=True` to enable. |
@@ -93,9 +93,9 @@ See full `export` details in the [Export](../modes/export.md) page.
## FAQ
-### How do I benchmark my YOLOv8 model's performance using Ultralytics?
+### How do I benchmark my YOLO11 model's performance using Ultralytics?
-Ultralytics YOLOv8 offers a Benchmark mode to assess your model's performance across different export formats. This mode provides insights into key metrics such as [mean Average Precision](https://www.ultralytics.com/glossary/mean-average-precision-map) (mAP50-95), accuracy, and inference time in milliseconds. To run benchmarks, you can use either Python or CLI commands. For example, to benchmark on a GPU:
+Ultralytics YOLO11 offers a Benchmark mode to assess your model's performance across different export formats. This mode provides insights into key metrics such as [mean Average Precision](https://www.ultralytics.com/glossary/mean-average-precision-map) (mAP50-95), accuracy, and inference time in milliseconds. To run benchmarks, you can use either Python or CLI commands. For example, to benchmark on a GPU:
!!! example
@@ -105,29 +105,29 @@ Ultralytics YOLOv8 offers a Benchmark mode to assess your model's performance ac
from ultralytics.utils.benchmarks import benchmark
# Benchmark on GPU
- benchmark(model="yolov8n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
+ benchmark(model="yolo11n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
```
=== "CLI"
```bash
- yolo benchmark model=yolov8n.pt data='coco8.yaml' imgsz=640 half=False device=0
+ yolo benchmark model=yolo11n.pt data='coco8.yaml' imgsz=640 half=False device=0
```
For more details on benchmark arguments, visit the [Arguments](#arguments) section.
-### What are the benefits of exporting YOLOv8 models to different formats?
+### What are the benefits of exporting YOLO11 models to different formats?
-Exporting YOLOv8 models to different formats such as ONNX, TensorRT, and OpenVINO allows you to optimize performance based on your deployment environment. For instance:
+Exporting YOLO11 models to different formats such as ONNX, TensorRT, and OpenVINO allows you to optimize performance based on your deployment environment. For instance:
- **ONNX:** Provides up to 3x CPU speedup.
- **TensorRT:** Offers up to 5x GPU speedup.
- **OpenVINO:** Specifically optimized for Intel hardware.
These formats enhance both the speed and accuracy of your models, making them more efficient for various real-world applications. Visit the [Export](../modes/export.md) page for complete details.
-### Why is benchmarking crucial in evaluating YOLOv8 models?
+### Why is benchmarking crucial in evaluating YOLO11 models?
-Benchmarking your YOLOv8 models is essential for several reasons:
+Benchmarking your YOLO11 models is essential for several reasons:
- **Informed Decisions:** Understand the trade-offs between speed and accuracy.
- **Resource Allocation:** Gauge the performance across different hardware options.
@@ -135,9 +135,9 @@ Benchmarking your YOLOv8 models is essential for several reasons:
- **Cost Efficiency:** Optimize hardware usage based on benchmark results.
Key metrics such as mAP50-95, Top-5 accuracy, and inference time help in making these evaluations. Refer to the [Key Metrics](#key-metrics-in-benchmark-mode) section for more information.
-### Which export formats are supported by YOLOv8, and what are their advantages?
+### Which export formats are supported by YOLO11, and what are their advantages?
-YOLOv8 supports a variety of export formats, each tailored for specific hardware and use cases:
+YOLO11 supports a variety of export formats, each tailored for specific hardware and use cases:
- **ONNX:** Best for CPU performance.
- **TensorRT:** Ideal for GPU efficiency.
@@ -145,11 +145,11 @@ YOLOv8 supports a variety of export formats, each tailored for specific hardware
- **CoreML & [TensorFlow](https://www.ultralytics.com/glossary/tensorflow):** Useful for iOS and general ML applications.
For a complete list of supported formats and their respective advantages, check out the [Supported Export Formats](#supported-export-formats) section.
-### What arguments can I use to fine-tune my YOLOv8 benchmarks?
+### What arguments can I use to fine-tune my YOLO11 benchmarks?
When running benchmarks, several arguments can be customized to suit specific needs:
-- **model:** Path to the model file (e.g., "yolov8n.pt").
+- **model:** Path to the model file (e.g., "yolo11n.pt").
- **data:** Path to a YAML file defining the dataset (e.g., "coco8.yaml").
- **imgsz:** The input image size, either as a single integer or a tuple.
- **half:** Enable FP16 inference for better performance.
diff --git a/docs/en/modes/export.md b/docs/en/modes/export.md
index 706dd91cdc..4be5bd5b90 100644
--- a/docs/en/modes/export.md
+++ b/docs/en/modes/export.md
@@ -1,7 +1,7 @@
---
comments: true
-description: Learn how to export your YOLOv8 model to various formats like ONNX, TensorRT, and CoreML. Achieve maximum compatibility and performance.
-keywords: YOLOv8, Model Export, ONNX, TensorRT, CoreML, Ultralytics, AI, Machine Learning, Inference, Deployment
+description: Learn how to export your YOLO11 model to various formats like ONNX, TensorRT, and CoreML. Achieve maximum compatibility and performance.
+keywords: YOLO11, Model Export, ONNX, TensorRT, CoreML, Ultralytics, AI, Machine Learning, Inference, Deployment
---
# Model Export with Ultralytics YOLO
@@ -10,7 +10,7 @@ keywords: YOLOv8, Model Export, ONNX, TensorRT, CoreML, Ultralytics, AI, Machine
## Introduction
-The ultimate goal of training a model is to deploy it for real-world applications. Export mode in Ultralytics YOLOv8 offers a versatile range of options for exporting your trained model to different formats, making it deployable across various platforms and devices. This comprehensive guide aims to walk you through the nuances of model exporting, showcasing how to achieve maximum compatibility and performance.
+The ultimate goal of training a model is to deploy it for real-world applications. Export mode in Ultralytics YOLO11 offers a versatile range of options for exporting your trained model to different formats, making it deployable across various platforms and devices. This comprehensive guide aims to walk you through the nuances of model exporting, showcasing how to achieve maximum compatibility and performance.
@@ -20,10 +20,10 @@ The ultimate goal of training a model is to deploy it for real-world application
allowfullscreen>
- Watch: How To Export Custom Trained Ultralytics YOLOv8 Model and Run Live Inference on Webcam.
+ Watch: How To Export Custom Trained Ultralytics YOLO Model and Run Live Inference on Webcam.
@@ -25,7 +25,7 @@ Ultralytics YOLOv8 is not just another object detection model; it's a versatile
### Modes at a Glance
-Understanding the different **modes** that Ultralytics YOLOv8 supports is critical to getting the most out of your models:
+Understanding the different **modes** that Ultralytics YOLO11 supports is critical to getting the most out of your models:
- **Train** mode: Fine-tune your model on custom or preloaded datasets.
- **Val** mode: A post-training checkpoint to validate model performance.
@@ -34,49 +34,49 @@ Understanding the different **modes** that Ultralytics YOLOv8 supports is critic
- **Track** mode: Extend your object detection model into real-time tracking applications.
- **Benchmark** mode: Analyze the speed and accuracy of your model in diverse deployment environments.
-This comprehensive guide aims to give you an overview and practical insights into each mode, helping you harness the full potential of YOLOv8.
+This comprehensive guide aims to give you an overview and practical insights into each mode, helping you harness the full potential of YOLO11.
## [Train](train.md)
-Train mode is used for training a YOLOv8 model on a custom dataset. In this mode, the model is trained using the specified dataset and hyperparameters. The training process involves optimizing the model's parameters so that it can accurately predict the classes and locations of objects in an image.
+Train mode is used for training a YOLO11 model on a custom dataset. In this mode, the model is trained using the specified dataset and hyperparameters. The training process involves optimizing the model's parameters so that it can accurately predict the classes and locations of objects in an image.
[Train Examples](train.md){ .md-button }
## [Val](val.md)
-Val mode is used for validating a YOLOv8 model after it has been trained. In this mode, the model is evaluated on a validation set to measure its accuracy and generalization performance. This mode can be used to tune the hyperparameters of the model to improve its performance.
+Val mode is used for validating a YOLO11 model after it has been trained. In this mode, the model is evaluated on a validation set to measure its accuracy and generalization performance. This mode can be used to tune the hyperparameters of the model to improve its performance.
[Val Examples](val.md){ .md-button }
## [Predict](predict.md)
-Predict mode is used for making predictions using a trained YOLOv8 model on new images or videos. In this mode, the model is loaded from a checkpoint file, and the user can provide images or videos to perform inference. The model predicts the classes and locations of objects in the input images or videos.
+Predict mode is used for making predictions using a trained YOLO11 model on new images or videos. In this mode, the model is loaded from a checkpoint file, and the user can provide images or videos to perform inference. The model predicts the classes and locations of objects in the input images or videos.
[Predict Examples](predict.md){ .md-button }
## [Export](export.md)
-Export mode is used for exporting a YOLOv8 model to a format that can be used for deployment. In this mode, the model is converted to a format that can be used by other software applications or hardware devices. This mode is useful when deploying the model to production environments.
+Export mode is used for exporting a YOLO11 model to a format that can be used for deployment. In this mode, the model is converted to a format that can be used by other software applications or hardware devices. This mode is useful when deploying the model to production environments.
[Export Examples](export.md){ .md-button }
## [Track](track.md)
-Track mode is used for tracking objects in real-time using a YOLOv8 model. In this mode, the model is loaded from a checkpoint file, and the user can provide a live video stream to perform real-time object tracking. This mode is useful for applications such as surveillance systems or self-driving cars.
+Track mode is used for tracking objects in real-time using a YOLO11 model. In this mode, the model is loaded from a checkpoint file, and the user can provide a live video stream to perform real-time object tracking. This mode is useful for applications such as surveillance systems or self-driving cars.
[Track Examples](track.md){ .md-button }
## [Benchmark](benchmark.md)
-Benchmark mode is used to profile the speed and accuracy of various export formats for YOLOv8. The benchmarks provide information on the size of the exported format, its `mAP50-95` metrics (for object detection, segmentation, and pose) or `accuracy_top5` metrics (for classification), and the inference time in milliseconds per image across various formats like ONNX, OpenVINO, TensorRT, and others. This information can help users choose the optimal export format for their specific use case based on their requirements for speed and accuracy.
+Benchmark mode is used to profile the speed and accuracy of various export formats for YOLO11. The benchmarks provide information on the size of the exported format, its `mAP50-95` metrics (for object detection, segmentation, and pose) or `accuracy_top5` metrics (for classification), and the inference time in milliseconds per image across various formats like ONNX, OpenVINO, TensorRT, and others. This information can help users choose the optimal export format for their specific use case based on their requirements for speed and accuracy.
[Benchmark Examples](benchmark.md){ .md-button }
## FAQ
-### How do I train a custom [object detection](https://www.ultralytics.com/glossary/object-detection) model with Ultralytics YOLOv8?
+### How do I train a custom [object detection](https://www.ultralytics.com/glossary/object-detection) model with Ultralytics YOLO11?
-Training a custom object detection model with Ultralytics YOLOv8 involves using the train mode. You need a dataset formatted in YOLO format, containing images and corresponding annotation files. Use the following command to start the training process:
+Training a custom object detection model with Ultralytics YOLO11 involves using the train mode. You need a dataset formatted in YOLO format, containing images and corresponding annotation files. Use the following command to start the training process:
!!! example
@@ -85,22 +85,25 @@ Training a custom object detection model with Ultralytics YOLOv8 involves using
```python
from ultralytics import YOLO
- # Train a custom model
- model = YOLO("yolov8n.pt")
+ # Load a pre-trained YOLO model (you can choose n, s, m, l, or x versions)
+ model = YOLO("yolo11n.pt")
+
+ # Start training on your custom dataset
model.train(data="path/to/dataset.yaml", epochs=100, imgsz=640)
```
=== "CLI"
```bash
+ # Train a YOLO model from the command line
yolo train data=path/to/dataset.yaml epochs=100 imgsz=640
```
For more detailed instructions, you can refer to the [Ultralytics Train Guide](../modes/train.md).
-### What metrics does Ultralytics YOLOv8 use to validate the model's performance?
+### What metrics does Ultralytics YOLO11 use to validate the model's performance?
-Ultralytics YOLOv8 uses various metrics during the validation process to assess model performance. These include:
+Ultralytics YOLO11 uses various metrics during the validation process to assess model performance. These include:
- **mAP (mean Average Precision)**: This evaluates the accuracy of object detection.
- **IOU (Intersection over Union)**: Measures the overlap between predicted and ground truth bounding boxes.
@@ -115,22 +118,25 @@ You can run the following command to start the validation:
```python
from ultralytics import YOLO
- # Validate the model
- model = YOLO("yolov8n.pt")
+ # Load a pre-trained or custom YOLO model
+ model = YOLO("yolo11n.pt")
+
+ # Run validation on your dataset
model.val(data="path/to/validation.yaml")
```
=== "CLI"
```bash
+ # Validate a YOLO model from the command line
yolo val data=path/to/validation.yaml
```
Refer to the [Validation Guide](../modes/val.md) for further details.
-### How can I export my YOLOv8 model for deployment?
+### How can I export my YOLO11 model for deployment?
-Ultralytics YOLOv8 offers export functionality to convert your trained model into various deployment formats such as ONNX, TensorRT, CoreML, and more. Use the following example to export your model:
+Ultralytics YOLO11 offers export functionality to convert your trained model into various deployment formats such as ONNX, TensorRT, CoreML, and more. Use the following example to export your model:
!!! example
@@ -139,22 +145,25 @@ Ultralytics YOLOv8 offers export functionality to convert your trained model int
```python
from ultralytics import YOLO
- # Export the model
- model = YOLO("yolov8n.pt")
+ # Load your trained YOLO model
+ model = YOLO("yolo11n.pt")
+
+ # Export the model to ONNX format (you can specify other formats as needed)
model.export(format="onnx")
```
=== "CLI"
```bash
- yolo export model=yolov8n.pt format=onnx
+ # Export a YOLO model to ONNX format from the command line
+ yolo export model=yolo11n.pt format=onnx
```
Detailed steps for each export format can be found in the [Export Guide](../modes/export.md).
-### What is the purpose of the benchmark mode in Ultralytics YOLOv8?
+### What is the purpose of the benchmark mode in Ultralytics YOLO11?
-Benchmark mode in Ultralytics YOLOv8 is used to analyze the speed and [accuracy](https://www.ultralytics.com/glossary/accuracy) of various export formats such as ONNX, TensorRT, and OpenVINO. It provides metrics like model size, `mAP50-95` for object detection, and inference time across different hardware setups, helping you choose the most suitable format for your deployment needs.
+Benchmark mode in Ultralytics YOLO11 is used to analyze the speed and [accuracy](https://www.ultralytics.com/glossary/accuracy) of various export formats such as ONNX, TensorRT, and OpenVINO. It provides metrics like model size, `mAP50-95` for object detection, and inference time across different hardware setups, helping you choose the most suitable format for your deployment needs.
!!! example
@@ -163,21 +172,24 @@ Benchmark mode in Ultralytics YOLOv8 is used to analyze the speed and [accuracy]
```python
from ultralytics.utils.benchmarks import benchmark
- # Benchmark on GPU
- benchmark(model="yolov8n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
+ # Run benchmark on GPU (device 0)
+ # You can adjust parameters like model, dataset, image size, and precision as needed
+ benchmark(model="yolo11n.pt", data="coco8.yaml", imgsz=640, half=False, device=0)
```
=== "CLI"
```bash
- yolo benchmark model=yolov8n.pt data='coco8.yaml' imgsz=640 half=False device=0
+ # Benchmark a YOLO model from the command line
+ # Adjust parameters as needed for your specific use case
+ yolo benchmark model=yolo11n.pt data='coco8.yaml' imgsz=640 half=False device=0
```
For more details, refer to the [Benchmark Guide](../modes/benchmark.md).
-### How can I perform real-time object tracking using Ultralytics YOLOv8?
+### How can I perform real-time object tracking using Ultralytics YOLO11?
-Real-time object tracking can be achieved using the track mode in Ultralytics YOLOv8. This mode extends object detection capabilities to track objects across video frames or live feeds. Use the following example to enable tracking:
+Real-time object tracking can be achieved using the track mode in Ultralytics YOLO11. This mode extends object detection capabilities to track objects across video frames or live feeds. Use the following example to enable tracking:
!!! example
@@ -186,14 +198,19 @@ Real-time object tracking can be achieved using the track mode in Ultralytics YO
```python
from ultralytics import YOLO
- # Track objects in a video
- model = YOLO("yolov8n.pt")
+ # Load a pre-trained YOLO model
+ model = YOLO("yolo11n.pt")
+
+ # Start tracking objects in a video
+ # You can also use live video streams or webcam input
model.track(source="path/to/video.mp4")
```
=== "CLI"
```bash
+ # Perform object tracking on a video from the command line
+ # You can specify different sources like webcam (0) or RTSP streams
yolo track source=path/to/video.mp4
```
diff --git a/docs/en/modes/predict.md b/docs/en/modes/predict.md
index 196d9e2028..cb8ca25e7e 100644
--- a/docs/en/modes/predict.md
+++ b/docs/en/modes/predict.md
@@ -1,7 +1,7 @@
---
comments: true
-description: Harness the power of Ultralytics YOLOv8 for real-time, high-speed inference on various data sources. Learn about predict mode, key features, and practical applications.
-keywords: Ultralytics, YOLOv8, model prediction, inference, predict mode, real-time inference, computer vision, machine learning, streaming, high performance
+description: Harness the power of Ultralytics YOLO11 for real-time, high-speed inference on various data sources. Learn about predict mode, key features, and practical applications.
+keywords: Ultralytics, YOLO11, model prediction, inference, predict mode, real-time inference, computer vision, machine learning, streaming, high performance
---
# Model Prediction with Ultralytics YOLO
@@ -10,7 +10,7 @@ keywords: Ultralytics, YOLOv8, model prediction, inference, predict mode, real-t
## Introduction
-In the world of [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) and [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv), the process of making sense out of visual data is called 'inference' or 'prediction'. Ultralytics YOLOv8 offers a powerful feature known as **predict mode** that is tailored for high-performance, real-time inference on a wide range of data sources.
+In the world of [machine learning](https://www.ultralytics.com/glossary/machine-learning-ml) and [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv), the process of making sense out of visual data is called 'inference' or 'prediction'. Ultralytics YOLO11 offers a powerful feature known as **predict mode** that is tailored for high-performance, real-time inference on a wide range of data sources.
@@ -20,7 +20,7 @@ In the world of [machine learning](https://www.ultralytics.com/glossary/machine-
allowfullscreen>
- Watch: How to Extract the Outputs from Ultralytics YOLOv8 Model for Custom Projects.
+ Watch: How to Extract the Outputs from Ultralytics YOLO Model for Custom Projects.
@@ -20,12 +20,12 @@ Training a [deep learning](https://www.ultralytics.com/glossary/deep-learning-dl
allowfullscreen>
- Watch: How to Train a YOLOv8 model on Your Custom Dataset in Google Colab.
+ Watch: How to Train a YOLO model on Your Custom Dataset in Google Colab.
@@ -25,7 +25,7 @@ Validation is a critical step in the [machine learning](https://www.ultralytics.
## Why Validate with Ultralytics YOLO?
-Here's why using YOLOv8's Val mode is advantageous:
+Here's why using YOLO11's Val mode is advantageous:
- **Precision:** Get accurate metrics like mAP50, mAP75, and mAP50-95 to comprehensively evaluate your model.
- **Convenience:** Utilize built-in features that remember training settings, simplifying the validation process.
@@ -34,7 +34,7 @@ Here's why using YOLOv8's Val mode is advantageous:
### Key Features of Val Mode
-These are the notable functionalities offered by YOLOv8's Val mode:
+These are the notable functionalities offered by YOLO11's Val mode:
- **Automated Settings:** Models remember their training configurations for straightforward validation.
- **Multi-Metric Support:** Evaluate your model based on a range of accuracy metrics.
@@ -43,11 +43,11 @@ These are the notable functionalities offered by YOLOv8's Val mode:
!!! tip
- * YOLOv8 models automatically remember their training settings, so you can validate a model at the same image size and on the original dataset easily with just `yolo val model=yolov8n.pt` or `model('yolov8n.pt').val()`
+ * YOLO11 models automatically remember their training settings, so you can validate a model at the same image size and on the original dataset easily with just `yolo val model=yolo11n.pt` or `model('yolo11n.pt').val()`
## Usage Examples
-Validate trained YOLOv8n model [accuracy](https://www.ultralytics.com/glossary/accuracy) on the COCO8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. See Arguments section below for a full list of export arguments.
+Validate trained YOLO11n model [accuracy](https://www.ultralytics.com/glossary/accuracy) on the COCO8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. See Arguments section below for a full list of export arguments.
!!! example
@@ -57,7 +57,7 @@ Validate trained YOLOv8n model [accuracy](https://www.ultralytics.com/glossary/a
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n.pt") # load an official model
+ model = YOLO("yolo11n.pt") # load an official model
model = YOLO("path/to/best.pt") # load a custom model
# Validate the model
@@ -71,7 +71,7 @@ Validate trained YOLOv8n model [accuracy](https://www.ultralytics.com/glossary/a
=== "CLI"
```bash
- yolo detect val model=yolov8n.pt # val official model
+ yolo detect val model=yolo11n.pt # val official model
yolo detect val model=path/to/best.pt # val custom model
```
@@ -95,7 +95,7 @@ The below examples showcase YOLO model validation with custom arguments in Pytho
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n.pt")
+ model = YOLO("yolo11n.pt")
# Customize validation settings
validation_results = model.val(data="coco8.yaml", imgsz=640, batch=16, conf=0.25, iou=0.6, device="0")
@@ -104,20 +104,20 @@ The below examples showcase YOLO model validation with custom arguments in Pytho
=== "CLI"
```bash
- yolo val model=yolov8n.pt data=coco8.yaml imgsz=640 batch=16 conf=0.25 iou=0.6 device=0
+ yolo val model=yolo11n.pt data=coco8.yaml imgsz=640 batch=16 conf=0.25 iou=0.6 device=0
```
## FAQ
-### How do I validate my YOLOv8 model with Ultralytics?
+### How do I validate my YOLO11 model with Ultralytics?
-To validate your YOLOv8 model, you can use the Val mode provided by Ultralytics. For example, using the Python API, you can load a model and run validation with:
+To validate your YOLO11 model, you can use the Val mode provided by Ultralytics. For example, using the Python API, you can load a model and run validation with:
```python
from ultralytics import YOLO
# Load a model
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
# Validate the model
metrics = model.val()
@@ -127,14 +127,14 @@ print(metrics.box.map) # map50-95
Alternatively, you can use the command-line interface (CLI):
```bash
-yolo val model=yolov8n.pt
+yolo val model=yolo11n.pt
```
For further customization, you can adjust various arguments like `imgsz`, `batch`, and `conf` in both Python and CLI modes. Check the [Arguments for YOLO Model Validation](#arguments-for-yolo-model-validation) section for the full list of parameters.
-### What metrics can I get from YOLOv8 model validation?
+### What metrics can I get from YOLO11 model validation?
-YOLOv8 model validation provides several key metrics to assess model performance. These include:
+YOLO11 model validation provides several key metrics to assess model performance. These include:
- mAP50 (mean Average Precision at IoU threshold 0.5)
- mAP75 (mean Average Precision at IoU threshold 0.75)
@@ -156,16 +156,16 @@ For a complete performance evaluation, it's crucial to review all these metrics.
Using Ultralytics YOLO for validation provides several advantages:
-- **[Precision](https://www.ultralytics.com/glossary/precision):** YOLOv8 offers accurate performance metrics including mAP50, mAP75, and mAP50-95.
+- **[Precision](https://www.ultralytics.com/glossary/precision):** YOLO11 offers accurate performance metrics including mAP50, mAP75, and mAP50-95.
- **Convenience:** The models remember their training settings, making validation straightforward.
- **Flexibility:** You can validate against the same or different datasets and image sizes.
- **Hyperparameter Tuning:** Validation metrics help in fine-tuning models for better performance.
These benefits ensure that your models are evaluated thoroughly and can be optimized for superior results. Learn more about these advantages in the [Why Validate with Ultralytics YOLO](#why-validate-with-ultralytics-yolo) section.
-### Can I validate my YOLOv8 model using a custom dataset?
+### Can I validate my YOLO11 model using a custom dataset?
-Yes, you can validate your YOLOv8 model using a [custom dataset](https://docs.ultralytics.com/datasets/). Specify the `data` argument with the path to your dataset configuration file. This file should include paths to the [validation data](https://www.ultralytics.com/glossary/validation-data), class names, and other relevant details.
+Yes, you can validate your YOLO11 model using a [custom dataset](https://docs.ultralytics.com/datasets/). Specify the `data` argument with the path to your dataset configuration file. This file should include paths to the [validation data](https://www.ultralytics.com/glossary/validation-data), class names, and other relevant details.
Example in Python:
@@ -173,7 +173,7 @@ Example in Python:
from ultralytics import YOLO
# Load a model
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
# Validate with a custom dataset
metrics = model.val(data="path/to/your/custom_dataset.yaml")
@@ -183,12 +183,12 @@ print(metrics.box.map) # map50-95
Example using CLI:
```bash
-yolo val model=yolov8n.pt data=path/to/your/custom_dataset.yaml
+yolo val model=yolo11n.pt data=path/to/your/custom_dataset.yaml
```
For more customizable options during validation, see the [Example Validation with Arguments](#example-validation-with-arguments) section.
-### How do I save validation results to a JSON file in YOLOv8?
+### How do I save validation results to a JSON file in YOLO11?
To save the validation results to a JSON file, you can set the `save_json` argument to `True` when running validation. This can be done in both the Python API and CLI.
@@ -198,7 +198,7 @@ Example in Python:
from ultralytics import YOLO
# Load a model
-model = YOLO("yolov8n.pt")
+model = YOLO("yolo11n.pt")
# Save validation results to JSON
metrics = model.val(save_json=True)
@@ -207,7 +207,7 @@ metrics = model.val(save_json=True)
Example using CLI:
```bash
-yolo val model=yolov8n.pt save_json=True
+yolo val model=yolo11n.pt save_json=True
```
This functionality is particularly useful for further analysis or integration with other tools. Check the [Arguments for YOLO Model Validation](#arguments-for-yolo-model-validation) for more details.
diff --git a/docs/en/solutions/index.md b/docs/en/solutions/index.md
index 52423c14f5..e5187ed8d4 100644
--- a/docs/en/solutions/index.md
+++ b/docs/en/solutions/index.md
@@ -1,12 +1,12 @@
---
comments: true
-description: Explore Ultralytics Solutions using YOLOv8 for object counting, blurring, security, and more. Enhance efficiency and solve real-world problems with cutting-edge AI.
-keywords: Ultralytics, YOLOv8, object counting, object blurring, security systems, AI solutions, real-time analysis, computer vision applications
+description: Explore Ultralytics Solutions using YOLO11 for object counting, blurring, security, and more. Enhance efficiency and solve real-world problems with cutting-edge AI.
+keywords: Ultralytics, YOLO11, object counting, object blurring, security systems, AI solutions, real-time analysis, computer vision applications
---
-# Ultralytics Solutions: Harness YOLOv8 to Solve Real-World Problems
+# Ultralytics Solutions: Harness YOLO11 to Solve Real-World Problems
-Ultralytics Solutions provide cutting-edge applications of YOLO models, offering real-world solutions like object counting, blurring, and security systems, enhancing efficiency and [accuracy](https://www.ultralytics.com/glossary/accuracy) in diverse industries. Discover the power of YOLOv8 for practical, impactful implementations.
+Ultralytics Solutions provide cutting-edge applications of YOLO models, offering real-world solutions like object counting, blurring, and security systems, enhancing efficiency and [accuracy](https://www.ultralytics.com/glossary/accuracy) in diverse industries. Discover the power of YOLO11 for practical, impactful implementations.
![Ultralytics Solutions Thumbnail](https://github.com/ultralytics/docs/releases/download/0/ultralytics-solutions-thumbnail.avif)
@@ -14,21 +14,21 @@ Ultralytics Solutions provide cutting-edge applications of YOLO models, offering
Here's our curated list of Ultralytics solutions that can be used to create awesome [computer vision](https://www.ultralytics.com/glossary/computer-vision-cv) projects.
-- [Object Counting](../guides/object-counting.md) 🚀 NEW: Learn to perform real-time object counting with YOLOv8. Gain the expertise to accurately count objects in live video streams.
-- [Object Cropping](../guides/object-cropping.md) 🚀 NEW: Master object cropping with YOLOv8 for precise extraction of objects from images and videos.
-- [Object Blurring](../guides/object-blurring.md) 🚀 NEW: Apply object blurring using YOLOv8 to protect privacy in image and video processing.
-- [Workouts Monitoring](../guides/workouts-monitoring.md) 🚀 NEW: Discover how to monitor workouts using YOLOv8. Learn to track and analyze various fitness routines in real time.
-- [Objects Counting in Regions](../guides/region-counting.md) 🚀 NEW: Count objects in specific regions using YOLOv8 for accurate detection in varied areas.
-- [Security Alarm System](../guides/security-alarm-system.md) 🚀 NEW: Create a security alarm system with YOLOv8 that triggers alerts upon detecting new objects. Customize the system to fit your specific needs.
+- [Object Counting](../guides/object-counting.md) 🚀 NEW: Learn to perform real-time object counting with YOLO11. Gain the expertise to accurately count objects in live video streams.
+- [Object Cropping](../guides/object-cropping.md) 🚀 NEW: Master object cropping with YOLO11 for precise extraction of objects from images and videos.
+- [Object Blurring](../guides/object-blurring.md) 🚀 NEW: Apply object blurring using YOLO11 to protect privacy in image and video processing.
+- [Workouts Monitoring](../guides/workouts-monitoring.md) 🚀 NEW: Discover how to monitor workouts using YOLO11. Learn to track and analyze various fitness routines in real time.
+- [Objects Counting in Regions](../guides/region-counting.md) 🚀 NEW: Count objects in specific regions using YOLO11 for accurate detection in varied areas.
+- [Security Alarm System](../guides/security-alarm-system.md) 🚀 NEW: Create a security alarm system with YOLO11 that triggers alerts upon detecting new objects. Customize the system to fit your specific needs.
- [Heatmaps](../guides/heatmaps.md) 🚀 NEW: Utilize detection heatmaps to visualize data intensity across a matrix, providing clear insights in computer vision tasks.
-- [Instance Segmentation with Object Tracking](../guides/instance-segmentation-and-tracking.md) 🚀 NEW: Implement [instance segmentation](https://www.ultralytics.com/glossary/instance-segmentation) and object tracking with YOLOv8 to achieve precise object boundaries and continuous monitoring.
+- [Instance Segmentation with Object Tracking](../guides/instance-segmentation-and-tracking.md) 🚀 NEW: Implement [instance segmentation](https://www.ultralytics.com/glossary/instance-segmentation) and object tracking with YOLO11 to achieve precise object boundaries and continuous monitoring.
- [VisionEye View Objects Mapping](../guides/vision-eye.md) 🚀 NEW: Develop systems that mimic human eye focus on specific objects, enhancing the computer's ability to discern and prioritize details.
-- [Speed Estimation](../guides/speed-estimation.md) 🚀 NEW: Estimate object speed using YOLOv8 and object tracking techniques, crucial for applications like autonomous vehicles and traffic monitoring.
-- [Distance Calculation](../guides/distance-calculation.md) 🚀 NEW: Calculate distances between objects using [bounding box](https://www.ultralytics.com/glossary/bounding-box) centroids in YOLOv8, essential for spatial analysis.
-- [Queue Management](../guides/queue-management.md) 🚀 NEW: Implement efficient queue management systems to minimize wait times and improve productivity using YOLOv8.
-- [Parking Management](../guides/parking-management.md) 🚀 NEW: Organize and direct vehicle flow in parking areas with YOLOv8, optimizing space utilization and user experience.
-- [Analytics](../guides/analytics.md) 📊 NEW: Conduct comprehensive data analysis to discover patterns and make informed decisions, leveraging YOLOv8 for descriptive, predictive, and prescriptive analytics.
-- [Live Inference with Streamlit](../guides/streamlit-live-inference.md) 🚀 NEW: Leverage the power of YOLOv8 for real-time [object detection](https://www.ultralytics.com/glossary/object-detection) directly through your web browser with a user-friendly Streamlit interface.
+- [Speed Estimation](../guides/speed-estimation.md) 🚀 NEW: Estimate object speed using YOLO11 and object tracking techniques, crucial for applications like autonomous vehicles and traffic monitoring.
+- [Distance Calculation](../guides/distance-calculation.md) 🚀 NEW: Calculate distances between objects using [bounding box](https://www.ultralytics.com/glossary/bounding-box) centroids in YOLO11, essential for spatial analysis.
+- [Queue Management](../guides/queue-management.md) 🚀 NEW: Implement efficient queue management systems to minimize wait times and improve productivity using YOLO11.
+- [Parking Management](../guides/parking-management.md) 🚀 NEW: Organize and direct vehicle flow in parking areas with YOLO11, optimizing space utilization and user experience.
+- [Analytics](../guides/analytics.md) 📊 NEW: Conduct comprehensive data analysis to discover patterns and make informed decisions, leveraging YOLO11 for descriptive, predictive, and prescriptive analytics.
+- [Live Inference with Streamlit](../guides/streamlit-live-inference.md) 🚀 NEW: Leverage the power of YOLO11 for real-time [object detection](https://www.ultralytics.com/glossary/object-detection) directly through your web browser with a user-friendly Streamlit interface.
## Contribute to Our Solutions
@@ -42,20 +42,20 @@ Let's work together to make the Ultralytics YOLO ecosystem more robust and versa
### How can I use Ultralytics YOLO for real-time object counting?
-Ultralytics YOLOv8 can be used for real-time object counting by leveraging its advanced object detection capabilities. You can follow our detailed guide on [Object Counting](../guides/object-counting.md) to set up YOLOv8 for live video stream analysis. Simply install YOLOv8, load your model, and process video frames to count objects dynamically.
+Ultralytics YOLO11 can be used for real-time object counting by leveraging its advanced object detection capabilities. You can follow our detailed guide on [Object Counting](../guides/object-counting.md) to set up YOLO11 for live video stream analysis. Simply install YOLO11, load your model, and process video frames to count objects dynamically.
### What are the benefits of using Ultralytics YOLO for security systems?
-Ultralytics YOLOv8 enhances security systems by offering real-time object detection and alert mechanisms. By employing YOLOv8, you can create a security alarm system that triggers alerts when new objects are detected in the surveillance area. Learn how to set up a [Security Alarm System](../guides/security-alarm-system.md) with YOLOv8 for robust security monitoring.
+Ultralytics YOLO11 enhances security systems by offering real-time object detection and alert mechanisms. By employing YOLO11, you can create a security alarm system that triggers alerts when new objects are detected in the surveillance area. Learn how to set up a [Security Alarm System](../guides/security-alarm-system.md) with YOLO11 for robust security monitoring.
### How can Ultralytics YOLO improve queue management systems?
-Ultralytics YOLOv8 can significantly improve queue management systems by accurately counting and tracking people in queues, thus helping to reduce wait times and optimize service efficiency. Follow our detailed guide on [Queue Management](../guides/queue-management.md) to learn how to implement YOLOv8 for effective queue monitoring and analysis.
+Ultralytics YOLO11 can significantly improve queue management systems by accurately counting and tracking people in queues, thus helping to reduce wait times and optimize service efficiency. Follow our detailed guide on [Queue Management](../guides/queue-management.md) to learn how to implement YOLO11 for effective queue monitoring and analysis.
### Can Ultralytics YOLO be used for workout monitoring?
-Yes, Ultralytics YOLOv8 can be effectively used for monitoring workouts by tracking and analyzing fitness routines in real-time. This allows for precise evaluation of exercise form and performance. Explore our guide on [Workouts Monitoring](../guides/workouts-monitoring.md) to learn how to set up an AI-powered workout monitoring system using YOLOv8.
+Yes, Ultralytics YOLO11 can be effectively used for monitoring workouts by tracking and analyzing fitness routines in real-time. This allows for precise evaluation of exercise form and performance. Explore our guide on [Workouts Monitoring](../guides/workouts-monitoring.md) to learn how to set up an AI-powered workout monitoring system using YOLO11.
### How does Ultralytics YOLO help in creating heatmaps for [data visualization](https://www.ultralytics.com/glossary/data-visualization)?
-Ultralytics YOLOv8 can generate heatmaps to visualize data intensity across a given area, highlighting regions of high activity or interest. This feature is particularly useful in understanding patterns and trends in various computer vision tasks. Learn more about creating and using [Heatmaps](../guides/heatmaps.md) with YOLOv8 for comprehensive data analysis and visualization.
+Ultralytics YOLO11 can generate heatmaps to visualize data intensity across a given area, highlighting regions of high activity or interest. This feature is particularly useful in understanding patterns and trends in various computer vision tasks. Learn more about creating and using [Heatmaps](../guides/heatmaps.md) with YOLO11 for comprehensive data analysis and visualization.
diff --git a/docs/en/tasks/classify.md b/docs/en/tasks/classify.md
index 62623a403f..7da8a3e7e7 100644
--- a/docs/en/tasks/classify.md
+++ b/docs/en/tasks/classify.md
@@ -1,8 +1,8 @@
---
comments: true
-description: Master image classification using YOLOv8. Learn to train, validate, predict, and export models efficiently.
-keywords: YOLOv8, image classification, AI, machine learning, pretrained models, ImageNet, model export, predict, train, validate
-model_name: yolov8n-cls
+description: Master image classification using YOLO11. Learn to train, validate, predict, and export models efficiently.
+keywords: YOLO11, image classification, AI, machine learning, pretrained models, ImageNet, model export, predict, train, validate
+model_name: yolo11n-cls
---
# Image Classification
@@ -26,28 +26,22 @@ The output of an image classifier is a single class label and a confidence score
!!! tip
- YOLOv8 Classify models use the `-cls` suffix, i.e. `yolov8n-cls.pt` and are pretrained on [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml).
+ YOLO11 Classify models use the `-cls` suffix, i.e. `yolo11n-cls.pt` and are pretrained on [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml).
-## [Models](https://github.com/ultralytics/ultralytics/tree/main/ultralytics/cfg/models/v8)
+## [Models](https://github.com/ultralytics/ultralytics/tree/main/ultralytics/cfg/models/11)
-YOLOv8 pretrained Classify models are shown here. Detect, Segment and Pose models are pretrained on the [COCO](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml) dataset, while Classify models are pretrained on the [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml) dataset.
+YOLO11 pretrained Classify models are shown here. Detect, Segment and Pose models are pretrained on the [COCO](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/coco.yaml) dataset, while Classify models are pretrained on the [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml) dataset.
[Models](https://github.com/ultralytics/ultralytics/tree/main/ultralytics/cfg/models) download automatically from the latest Ultralytics [release](https://github.com/ultralytics/assets/releases) on first use.
-| Model | size
(pixels) | acc
top1 | acc
top5 | Speed
CPU ONNX
(ms) | Speed
A100 TensorRT
(ms) | params
(M) | FLOPs
(B) at 640 |
-| -------------------------------------------------------------------------------------------- | --------------------- | ---------------- | ---------------- | ------------------------------ | ----------------------------------- | ------------------ | ------------------------ |
-| [YOLOv8n-cls](https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8n-cls.pt) | 224 | 69.0 | 88.3 | 12.9 | 0.31 | 2.7 | 4.3 |
-| [YOLOv8s-cls](https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8s-cls.pt) | 224 | 73.8 | 91.7 | 23.4 | 0.35 | 6.4 | 13.5 |
-| [YOLOv8m-cls](https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8m-cls.pt) | 224 | 76.8 | 93.5 | 85.4 | 0.62 | 17.0 | 42.7 |
-| [YOLOv8l-cls](https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8l-cls.pt) | 224 | 76.8 | 93.5 | 163.0 | 0.87 | 37.5 | 99.7 |
-| [YOLOv8x-cls](https://github.com/ultralytics/assets/releases/download/v8.2.0/yolov8x-cls.pt) | 224 | 79.0 | 94.6 | 232.0 | 1.01 | 57.4 | 154.8 |
+{% include "macros/yolo-cls-perf.md" %}
- **acc** values are model accuracies on the [ImageNet](https://www.image-net.org/) dataset validation set.
Reproduce by `yolo val classify data=path/to/ImageNet device=0`
- **Speed** averaged over ImageNet val images using an [Amazon EC2 P4d](https://aws.amazon.com/ec2/instance-types/p4/) instance.
Reproduce by `yolo val classify data=path/to/ImageNet batch=1 device=0|cpu`
## Train
-Train YOLOv8n-cls on the MNIST160 dataset for 100 [epochs](https://www.ultralytics.com/glossary/epoch) at image size 64. For a full list of available arguments see the [Configuration](../usage/cfg.md) page.
+Train YOLO11n-cls on the MNIST160 dataset for 100 [epochs](https://www.ultralytics.com/glossary/epoch) at image size 64. For a full list of available arguments see the [Configuration](../usage/cfg.md) page.
!!! example
@@ -57,9 +51,9 @@ Train YOLOv8n-cls on the MNIST160 dataset for 100 [epochs](https://www.ultralyti
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.yaml") # build a new model from YAML
- model = YOLO("yolov8n-cls.pt") # load a pretrained model (recommended for training)
- model = YOLO("yolov8n-cls.yaml").load("yolov8n-cls.pt") # build from YAML and transfer weights
+ model = YOLO("yolo11n-cls.yaml") # build a new model from YAML
+ model = YOLO("yolo11n-cls.pt") # load a pretrained model (recommended for training)
+ model = YOLO("yolo11n-cls.yaml").load("yolo11n-cls.pt") # build from YAML and transfer weights
# Train the model
results = model.train(data="mnist160", epochs=100, imgsz=64)
@@ -69,13 +63,13 @@ Train YOLOv8n-cls on the MNIST160 dataset for 100 [epochs](https://www.ultralyti
```bash
# Build a new model from YAML and start training from scratch
- yolo classify train data=mnist160 model=yolov8n-cls.yaml epochs=100 imgsz=64
+ yolo classify train data=mnist160 model=yolo11n-cls.yaml epochs=100 imgsz=64
# Start training from a pretrained *.pt model
- yolo classify train data=mnist160 model=yolov8n-cls.pt epochs=100 imgsz=64
+ yolo classify train data=mnist160 model=yolo11n-cls.pt epochs=100 imgsz=64
# Build a new model from YAML, transfer pretrained weights to it and start training
- yolo classify train data=mnist160 model=yolov8n-cls.yaml pretrained=yolov8n-cls.pt epochs=100 imgsz=64
+ yolo classify train data=mnist160 model=yolo11n-cls.yaml pretrained=yolo11n-cls.pt epochs=100 imgsz=64
```
### Dataset format
@@ -84,7 +78,7 @@ YOLO classification dataset format can be found in detail in the [Dataset Guide]
## Val
-Validate trained YOLOv8n-cls model [accuracy](https://www.ultralytics.com/glossary/accuracy) on the MNIST160 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes.
+Validate trained YOLO11n-cls model [accuracy](https://www.ultralytics.com/glossary/accuracy) on the MNIST160 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes.
!!! example
@@ -94,7 +88,7 @@ Validate trained YOLOv8n-cls model [accuracy](https://www.ultralytics.com/glossa
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load an official model
+ model = YOLO("yolo11n-cls.pt") # load an official model
model = YOLO("path/to/best.pt") # load a custom model
# Validate the model
@@ -106,13 +100,13 @@ Validate trained YOLOv8n-cls model [accuracy](https://www.ultralytics.com/glossa
=== "CLI"
```bash
- yolo classify val model=yolov8n-cls.pt # val official model
+ yolo classify val model=yolo11n-cls.pt # val official model
yolo classify val model=path/to/best.pt # val custom model
```
## Predict
-Use a trained YOLOv8n-cls model to run predictions on images.
+Use a trained YOLO11n-cls model to run predictions on images.
!!! example
@@ -122,7 +116,7 @@ Use a trained YOLOv8n-cls model to run predictions on images.
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load an official model
+ model = YOLO("yolo11n-cls.pt") # load an official model
model = YOLO("path/to/best.pt") # load a custom model
# Predict with the model
@@ -132,7 +126,7 @@ Use a trained YOLOv8n-cls model to run predictions on images.
=== "CLI"
```bash
- yolo classify predict model=yolov8n-cls.pt source='https://ultralytics.com/images/bus.jpg' # predict with official model
+ yolo classify predict model=yolo11n-cls.pt source='https://ultralytics.com/images/bus.jpg' # predict with official model
yolo classify predict model=path/to/best.pt source='https://ultralytics.com/images/bus.jpg' # predict with custom model
```
@@ -140,7 +134,7 @@ See full `predict` mode details in the [Predict](../modes/predict.md) page.
## Export
-Export a YOLOv8n-cls model to a different format like ONNX, CoreML, etc.
+Export a YOLO11n-cls model to a different format like ONNX, CoreML, etc.
!!! example
@@ -150,7 +144,7 @@ Export a YOLOv8n-cls model to a different format like ONNX, CoreML, etc.
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load an official model
+ model = YOLO("yolo11n-cls.pt") # load an official model
model = YOLO("path/to/best.pt") # load a custom trained model
# Export the model
@@ -160,11 +154,11 @@ Export a YOLOv8n-cls model to a different format like ONNX, CoreML, etc.
=== "CLI"
```bash
- yolo export model=yolov8n-cls.pt format=onnx # export official model
+ yolo export model=yolo11n-cls.pt format=onnx # export official model
yolo export model=path/to/best.pt format=onnx # export custom trained model
```
-Available YOLOv8-cls export formats are in the table below. You can export to any format using the `format` argument, i.e. `format='onnx'` or `format='engine'`. You can predict or validate directly on exported models, i.e. `yolo predict model=yolov8n-cls.onnx`. Usage examples are shown for your model after export completes.
+Available YOLO11-cls export formats are in the table below. You can export to any format using the `format` argument, i.e. `format='onnx'` or `format='engine'`. You can predict or validate directly on exported models, i.e. `yolo predict model=yolo11n-cls.onnx`. Usage examples are shown for your model after export completes.
{% include "macros/export-table.md" %}
@@ -172,13 +166,13 @@ See full `export` details in the [Export](../modes/export.md) page.
## FAQ
-### What is the purpose of YOLOv8 in image classification?
+### What is the purpose of YOLO11 in image classification?
-YOLOv8 models, such as `yolov8n-cls.pt`, are designed for efficient image classification. They assign a single class label to an entire image along with a confidence score. This is particularly useful for applications where knowing the specific class of an image is sufficient, rather than identifying the location or shape of objects within the image.
+YOLO11 models, such as `yolo11n-cls.pt`, are designed for efficient image classification. They assign a single class label to an entire image along with a confidence score. This is particularly useful for applications where knowing the specific class of an image is sufficient, rather than identifying the location or shape of objects within the image.
-### How do I train a YOLOv8 model for image classification?
+### How do I train a YOLO11 model for image classification?
-To train a YOLOv8 model, you can use either Python or CLI commands. For example, to train a `yolov8n-cls` model on the MNIST160 dataset for 100 epochs at an image size of 64:
+To train a YOLO11 model, you can use either Python or CLI commands. For example, to train a `yolo11n-cls` model on the MNIST160 dataset for 100 epochs at an image size of 64:
!!! example
@@ -188,7 +182,7 @@ To train a YOLOv8 model, you can use either Python or CLI commands. For example,
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load a pretrained model (recommended for training)
+ model = YOLO("yolo11n-cls.pt") # load a pretrained model (recommended for training)
# Train the model
results = model.train(data="mnist160", epochs=100, imgsz=64)
@@ -197,18 +191,18 @@ To train a YOLOv8 model, you can use either Python or CLI commands. For example,
=== "CLI"
```bash
- yolo classify train data=mnist160 model=yolov8n-cls.pt epochs=100 imgsz=64
+ yolo classify train data=mnist160 model=yolo11n-cls.pt epochs=100 imgsz=64
```
For more configuration options, visit the [Configuration](../usage/cfg.md) page.
-### Where can I find pretrained YOLOv8 classification models?
+### Where can I find pretrained YOLO11 classification models?
-Pretrained YOLOv8 classification models can be found in the [Models](https://github.com/ultralytics/ultralytics/tree/main/ultralytics/cfg/models/v8) section. Models like `yolov8n-cls.pt`, `yolov8s-cls.pt`, `yolov8m-cls.pt`, etc., are pretrained on the [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml) dataset and can be easily downloaded and used for various image classification tasks.
+Pretrained YOLO11 classification models can be found in the [Models](https://github.com/ultralytics/ultralytics/tree/main/ultralytics/cfg/models/11) section. Models like `yolo11n-cls.pt`, `yolo11s-cls.pt`, `yolo11m-cls.pt`, etc., are pretrained on the [ImageNet](https://github.com/ultralytics/ultralytics/blob/main/ultralytics/cfg/datasets/ImageNet.yaml) dataset and can be easily downloaded and used for various image classification tasks.
-### How can I export a trained YOLOv8 model to different formats?
+### How can I export a trained YOLO11 model to different formats?
-You can export a trained YOLOv8 model to various formats using Python or CLI commands. For instance, to export a model to ONNX format:
+You can export a trained YOLO11 model to various formats using Python or CLI commands. For instance, to export a model to ONNX format:
!!! example
@@ -218,7 +212,7 @@ You can export a trained YOLOv8 model to various formats using Python or CLI com
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load the trained model
+ model = YOLO("yolo11n-cls.pt") # load the trained model
# Export the model to ONNX
model.export(format="onnx")
@@ -227,12 +221,12 @@ You can export a trained YOLOv8 model to various formats using Python or CLI com
=== "CLI"
```bash
- yolo export model=yolov8n-cls.pt format=onnx # export the trained model to ONNX format
+ yolo export model=yolo11n-cls.pt format=onnx # export the trained model to ONNX format
```
For detailed export options, refer to the [Export](../modes/export.md) page.
-### How do I validate a trained YOLOv8 classification model?
+### How do I validate a trained YOLO11 classification model?
To validate a trained model's accuracy on a dataset like MNIST160, you can use the following Python or CLI commands:
@@ -244,7 +238,7 @@ To validate a trained model's accuracy on a dataset like MNIST160, you can use t
from ultralytics import YOLO
# Load a model
- model = YOLO("yolov8n-cls.pt") # load the trained model
+ model = YOLO("yolo11n-cls.pt") # load the trained model
# Validate the model
metrics = model.val() # no arguments needed, uses the dataset and settings from training
@@ -255,7 +249,7 @@ To validate a trained model's accuracy on a dataset like MNIST160, you can use t
=== "CLI"
```bash
- yolo classify val model=yolov8n-cls.pt # validate the trained model
+ yolo classify val model=yolo11n-cls.pt # validate the trained model
```
For more information, visit the [Validate](#val) section.
diff --git a/docs/en/tasks/detect.md b/docs/en/tasks/detect.md
index a9c54889bc..68c53f075b 100644
--- a/docs/en/tasks/detect.md
+++ b/docs/en/tasks/detect.md
@@ -1,7 +1,7 @@
---
comments: true
-description: Learn about object detection with YOLOv8. Explore pretrained models, training, validation, prediction, and export details for efficient object recognition.
-keywords: object detection, YOLOv8, pretrained models, training, validation, prediction, export, machine learning, computer vision
+description: Learn about object detection with YOLO11. Explore pretrained models, training, validation, prediction, and export details for efficient object recognition.
+keywords: object detection, YOLO11, pretrained models, training, validation, prediction, export, machine learning, computer vision
---
# Object Detection
@@ -20,33 +20,27 @@ The output of an object detector is a set of bounding boxes that enclose the obj
allowfullscreen>
- Watch: Object Detection with Pre-trained Ultralytics YOLOv8 Model.
+ Watch: Object Detection with Pre-trained Ultralytics YOLO Model.
@@ -19,48 +19,48 @@ YOLOv8 is an AI framework that supports multiple [computer vision](https://www.u
allowfullscreen>
- Watch: Explore Ultralytics YOLO Tasks: [Object Detection](https://www.ultralytics.com/glossary/object-detection), Segmentation, OBB, Tracking, and Pose Estimation.
+ Watch: Explore Ultralytics YOLO Tasks: Object Detection, Segmentation, OBB, Tracking, and Pose Estimation.
@@ -27,7 +27,7 @@ The output of an oriented object detector is a set of rotated bounding boxes tha
allowfullscreen>
- Watch: Object Detection using Ultralytics YOLOv8 Oriented Bounding Boxes (YOLOv8-OBB)
+ Watch: Object Detection using Ultralytics YOLO Oriented Bounding Boxes (YOLO-OBB)
@@ -94,7 +88,7 @@ Train YOLOv8n-obb on the `dota8.yaml` dataset for 100 [epochs](https://www.ultra
allowfullscreen>
- Watch: How to Train Ultralytics YOLOv8-OBB (Oriented Bounding Boxes) Models on DOTA Dataset using Ultralytics HUB
+ Watch: How to Train Ultralytics YOLO-OBB (Oriented Bounding Boxes) Models on DOTA Dataset using Ultralytics HUB
@@ -16,7 +16,7 @@ Welcome to the YOLOv8 Python Usage documentation! This guide is designed to help
allowfullscreen>
- Watch: Mastering Ultralytics YOLOv8: Python
+ Watch: Mastering Ultralytics YOLO11: Python