From 887b46216cb5b987a178e2c0e010d58fa4f15461 Mon Sep 17 00:00:00 2001 From: Glenn Jocher Date: Sun, 15 Sep 2024 00:27:46 +0200 Subject: [PATCH] Docs improvements and redirect fixes (#16287) Signed-off-by: UltralyticsAssistant Co-authored-by: UltralyticsAssistant --- CONTRIBUTING.md | 10 +++++----- README.md | 14 +++++++------- README.zh-CN.md | 14 +++++++------- docs/README.md | 4 ++-- docs/coming_soon_template.md | 2 +- docs/en/datasets/index.md | 4 ++-- docs/en/datasets/segment/coco8-seg.md | 2 +- docs/en/datasets/segment/package-seg.md | 2 +- docs/en/guides/object-blurring.md | 4 ++-- docs/en/guides/object-counting.md | 2 +- ...imizing-openvino-latency-vs-throughput-modes.md | 8 ++++---- docs/en/guides/raspberry-pi.md | 4 ++-- docs/en/guides/security-alarm-system.md | 2 +- docs/en/guides/streamlit-live-inference.md | 4 ++-- docs/en/guides/triton-inference-server.md | 4 ++-- docs/en/help/contributing.md | 10 +++++----- docs/en/hub/datasets.md | 2 +- docs/en/hub/index.md | 4 ++-- docs/en/hub/inference-api.md | 2 +- docs/en/hub/integrations.md | 2 +- docs/en/hub/models.md | 4 ++-- docs/en/hub/quickstart.md | 6 +++--- docs/en/integrations/edge-tpu.md | 2 +- docs/en/modes/export.md | 10 +++++----- docs/en/modes/index.md | 2 +- docs/en/modes/train.md | 6 +++--- docs/en/modes/val.md | 4 ++-- docs/en/tasks/classify.md | 2 +- docs/en/tasks/detect.md | 2 +- docs/en/tasks/obb.md | 3 +-- docs/en/tasks/pose.md | 3 +-- docs/en/tasks/segment.md | 3 +-- docs/en/usage/cli.md | 2 +- docs/en/yolov5/tutorials/train_custom_data.md | 2 +- examples/README.md | 2 +- pyproject.toml | 4 ++-- ultralytics/cfg/models/README.md | 8 ++++---- ultralytics/trackers/README.md | 2 +- 38 files changed, 82 insertions(+), 85 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index d884e43b4a..b3dbfe1695 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -25,7 +25,7 @@ Welcome! We're thrilled that you're considering contributing to our [Ultralytics ## Code of Conduct -To ensure a welcoming and inclusive environment for everyone, all contributors must adhere to our [Code of Conduct](https://docs.ultralytics.com/help/code_of_conduct). Respect, kindness, and professionalism are at the heart of our community. +To ensure a welcoming and inclusive environment for everyone, all contributors must adhere to our [Code of Conduct](https://docs.ultralytics.com/help/code_of_conduct/). Respect, kindness, and professionalism are at the heart of our community. ## Contributing via Pull Requests @@ -45,7 +45,7 @@ We greatly appreciate contributions in the form of pull requests. To make the re ### CLA Signing -Before we can merge your pull request, you must sign our [Contributor License Agreement (CLA)](https://docs.ultralytics.com/help/CLA). This legal agreement ensures that your contributions are properly licensed, allowing the project to continue being distributed under the AGPL-3.0 license. +Before we can merge your pull request, you must sign our [Contributor License Agreement (CLA)](https://docs.ultralytics.com/help/CLA/). This legal agreement ensures that your contributions are properly licensed, allowing the project to continue being distributed under the AGPL-3.0 license. After submitting your pull request, the CLA bot will guide you through the signing process. To sign the CLA, simply add a comment in your PR stating: @@ -117,11 +117,11 @@ def example_small_function(arg1: int, arg2: int = 4) -> bool: ### GitHub Actions CI Tests -All pull requests must pass the GitHub Actions [Continuous Integration](https://docs.ultralytics.com/help/CI) (CI) tests before they can be merged. These tests include linting, unit tests, and other checks to ensure that your changes meet the project's quality standards. Review the CI output and address any issues that arise. +All pull requests must pass the GitHub Actions [Continuous Integration](https://docs.ultralytics.com/help/CI/) (CI) tests before they can be merged. These tests include linting, unit tests, and other checks to ensure that your changes meet the project's quality standards. Review the CI output and address any issues that arise. ## Reporting Bugs -We highly value bug reports as they help us maintain the quality of our projects. When reporting a bug, please provide a [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example)—a simple, clear code example that consistently reproduces the issue. This allows us to quickly identify and resolve the problem. +We highly value bug reports as they help us maintain the quality of our projects. When reporting a bug, please provide a [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example/)—a simple, clear code example that consistently reproduces the issue. This allows us to quickly identify and resolve the problem. ## License @@ -163,4 +163,4 @@ the project's quality standards. Review the CI output and fix any issues. For de ### How do I report a bug in Ultralytics YOLO repositories? -To report a bug, provide a clear and concise [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example) along with your bug report. This helps developers quickly identify and fix the issue. Ensure your example is minimal yet sufficient to replicate the problem. For more detailed steps on reporting bugs, refer to the [Reporting Bugs](#reporting-bugs) section. +To report a bug, provide a clear and concise [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example/) along with your bug report. This helps developers quickly identify and fix the issue. Ensure your example is minimal yet sufficient to replicate the problem. For more detailed steps on reporting bugs, refer to the [Reporting Bugs](#reporting-bugs) section. diff --git a/README.md b/README.md index 1cc1686e67..e13ef56450 100644 --- a/README.md +++ b/README.md @@ -47,7 +47,7 @@ To request an Enterprise License please complete the form at [Ultralytics Licens ##
Documentation
-See below for a quickstart installation and usage example, and see the [YOLOv8 Docs](https://docs.ultralytics.com) for full documentation on training, validation, prediction and deployment. +See below for a quickstart installation and usage example, and see the [YOLOv8 Docs](https://docs.ultralytics.com/) for full documentation on training, validation, prediction and deployment.
Install @@ -60,7 +60,7 @@ Pip install the ultralytics package including all [requirements](https://github. pip install ultralytics ``` -For alternative installation methods including [Conda](https://anaconda.org/conda-forge/ultralytics), [Docker](https://hub.docker.com/r/ultralytics/ultralytics), and Git, please refer to the [Quickstart Guide](https://docs.ultralytics.com/quickstart). +For alternative installation methods including [Conda](https://anaconda.org/conda-forge/ultralytics), [Docker](https://hub.docker.com/r/ultralytics/ultralytics), and Git, please refer to the [Quickstart Guide](https://docs.ultralytics.com/quickstart/). [![Conda Version](https://img.shields.io/conda/vn/conda-forge/ultralytics?logo=condaforge)](https://anaconda.org/conda-forge/ultralytics) [![Docker Image Version](https://img.shields.io/docker/v/ultralytics/ultralytics?sort=semver&logo=docker)](https://hub.docker.com/r/ultralytics/ultralytics) @@ -77,7 +77,7 @@ YOLOv8 may be used directly in the Command Line Interface (CLI) with a `yolo` co yolo predict model=yolov8n.pt source='https://ultralytics.com/images/bus.jpg' ``` -`yolo` can be used for a variety of tasks and modes and accepts additional arguments, i.e. `imgsz=640`. See the YOLOv8 [CLI Docs](https://docs.ultralytics.com/usage/cli) for examples. +`yolo` can be used for a variety of tasks and modes and accepts additional arguments, i.e. `imgsz=640`. See the YOLOv8 [CLI Docs](https://docs.ultralytics.com/usage/cli/) for examples. ### Python @@ -97,7 +97,7 @@ results = model("https://ultralytics.com/images/bus.jpg") # predict on an image path = model.export(format="onnx") # export the model to ONNX format ``` -See YOLOv8 [Python Docs](https://docs.ultralytics.com/usage/python) for more examples. +See YOLOv8 [Python Docs](https://docs.ultralytics.com/usage/python/) for more examples.
@@ -116,7 +116,7 @@ Ultralytics provides interactive notebooks for YOLOv8, covering training, valida ##
Models
-YOLOv8 [Detect](https://docs.ultralytics.com/tasks/detect), [Segment](https://docs.ultralytics.com/tasks/segment) and [Pose](https://docs.ultralytics.com/tasks/pose) models pretrained on the [COCO](https://docs.ultralytics.com/datasets/detect/coco) dataset are available here, as well as YOLOv8 [Classify](https://docs.ultralytics.com/tasks/classify) models pretrained on the [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet) dataset. [Track](https://docs.ultralytics.com/modes/track) mode is available for all Detect, Segment and Pose models. +YOLOv8 [Detect](https://docs.ultralytics.com/tasks/detect/), [Segment](https://docs.ultralytics.com/tasks/segment/) and [Pose](https://docs.ultralytics.com/tasks/pose/) models pretrained on the [COCO](https://docs.ultralytics.com/datasets/detect/coco/) dataset are available here, as well as YOLOv8 [Classify](https://docs.ultralytics.com/tasks/classify/) models pretrained on the [ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/) dataset. [Track](https://docs.ultralytics.com/modes/track/) mode is available for all Detect, Segment and Pose models. Ultralytics YOLO supported tasks @@ -227,7 +227,7 @@ See [Classification Docs](https://docs.ultralytics.com/tasks/classify/) for usag ##
Integrations
-Our key integrations with leading AI platforms extend the functionality of Ultralytics' offerings, enhancing tasks like dataset labeling, training, visualization, and model management. Discover how Ultralytics, in collaboration with [Roboflow](https://roboflow.com/?ref=ultralytics), ClearML, [Comet](https://bit.ly/yolov8-readme-comet), Neural Magic and [OpenVINO](https://docs.ultralytics.com/integrations/openvino), can optimize your AI workflow. +Our key integrations with leading AI platforms extend the functionality of Ultralytics' offerings, enhancing tasks like dataset labeling, training, visualization, and model management. Discover how Ultralytics, in collaboration with [Roboflow](https://roboflow.com/?ref=ultralytics), ClearML, [Comet](https://bit.ly/yolov8-readme-comet), Neural Magic and [OpenVINO](https://docs.ultralytics.com/integrations/openvino/), can optimize your AI workflow.
@@ -262,7 +262,7 @@ Experience seamless AI with [Ultralytics HUB](https://www.ultralytics.com/hub) ##
Contribute
-We love your input! YOLOv5 and YOLOv8 would not be possible without help from our community. Please see our [Contributing Guide](https://docs.ultralytics.com/help/contributing) to get started, and fill out our [Survey](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey) to send us feedback on your experience. Thank you 🙏 to all our contributors! +We love your input! YOLOv5 and YOLOv8 would not be possible without help from our community. Please see our [Contributing Guide](https://docs.ultralytics.com/help/contributing/) to get started, and fill out our [Survey](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey) to send us feedback on your experience. Thank you 🙏 to all our contributors! diff --git a/README.zh-CN.md b/README.zh-CN.md index 79de5de2bc..4319c0dff8 100644 --- a/README.zh-CN.md +++ b/README.zh-CN.md @@ -49,7 +49,7 @@ ##
文档
-请参阅下面的快速安装和使用示例,以及 [YOLOv8 文档](https://docs.ultralytics.com) 上有关训练、验证、预测和部署的完整文档。 +请参阅下面的快速安装和使用示例,以及 [YOLOv8 文档](https://docs.ultralytics.com/) 上有关训练、验证、预测和部署的完整文档。
安装 @@ -62,7 +62,7 @@ pip install ultralytics ``` -如需使用包括[Conda](https://anaconda.org/conda-forge/ultralytics),[Docker](https://hub.docker.com/r/ultralytics/ultralytics)和Git在内的其他安装方法,请参考[快速入门指南](https://docs.ultralytics.com/quickstart)。 +如需使用包括[Conda](https://anaconda.org/conda-forge/ultralytics),[Docker](https://hub.docker.com/r/ultralytics/ultralytics)和Git在内的其他安装方法,请参考[快速入门指南](https://docs.ultralytics.com/quickstart/)。 [![Conda Version](https://img.shields.io/conda/vn/conda-forge/ultralytics?logo=condaforge)](https://anaconda.org/conda-forge/ultralytics) [![Docker Image Version](https://img.shields.io/docker/v/ultralytics/ultralytics?sort=semver&logo=docker)](https://hub.docker.com/r/ultralytics/ultralytics) @@ -79,7 +79,7 @@ YOLOv8 可以在命令行界面(CLI)中直接使用,只需输入 `yolo` yolo predict model=yolov8n.pt source='https://ultralytics.com/images/bus.jpg' ``` -`yolo` 可用于各种任务和模式,并接受其他参数,例如 `imgsz=640`。查看 YOLOv8 [CLI 文档](https://docs.ultralytics.com/usage/cli)以获取示例。 +`yolo` 可用于各种任务和模式,并接受其他参数,例如 `imgsz=640`。查看 YOLOv8 [CLI 文档](https://docs.ultralytics.com/usage/cli/)以获取示例。 ### Python @@ -99,7 +99,7 @@ results = model("https://ultralytics.com/images/bus.jpg") # 对图像进行预 success = model.export(format="onnx") # 将模型导出为 ONNX 格式 ``` -查看 YOLOv8 [Python 文档](https://docs.ultralytics.com/usage/python)以获取更多示例。 +查看 YOLOv8 [Python 文档](https://docs.ultralytics.com/usage/python/)以获取更多示例。
@@ -118,7 +118,7 @@ Ultralytics 提供了 YOLOv8 的交互式笔记本,涵盖训练、验证、跟 ##
模型
-在[COCO](https://docs.ultralytics.com/datasets/detect/coco)数据集上预训练的YOLOv8 [检测](https://docs.ultralytics.com/tasks/detect),[分割](https://docs.ultralytics.com/tasks/segment)和[姿态](https://docs.ultralytics.com/tasks/pose)模型可以在这里找到,以及在[ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet)数据集上预训练的YOLOv8 [分类](https://docs.ultralytics.com/tasks/classify)模型。所有的检测,分割和姿态模型都支持[追踪](https://docs.ultralytics.com/modes/track)模式。 +在[COCO](https://docs.ultralytics.com/datasets/detect/coco/)数据集上预训练的YOLOv8 [检测](https://docs.ultralytics.com/tasks/detect/),[分割](https://docs.ultralytics.com/tasks/segment/)和[姿态](https://docs.ultralytics.com/tasks/pose/)模型可以在这里找到,以及在[ImageNet](https://docs.ultralytics.com/datasets/classify/imagenet/)数据集上预训练的YOLOv8 [分类](https://docs.ultralytics.com/tasks/classify/)模型。所有的检测,分割和姿态模型都支持[追踪](https://docs.ultralytics.com/modes/track/)模式。 Ultralytics YOLO supported tasks @@ -229,7 +229,7 @@ Ultralytics 提供了 YOLOv8 的交互式笔记本,涵盖训练、验证、跟 ##
集成
-我们与领先的AI平台的关键整合扩展了Ultralytics产品的功能,增强了数据集标签化、训练、可视化和模型管理等任务。探索Ultralytics如何与[Roboflow](https://roboflow.com/?ref=ultralytics)、ClearML、[Comet](https://bit.ly/yolov8-readme-comet)、Neural Magic以及[OpenVINO](https://docs.ultralytics.com/integrations/openvino)合作,优化您的AI工作流程。 +我们与领先的AI平台的关键整合扩展了Ultralytics产品的功能,增强了数据集标签化、训练、可视化和模型管理等任务。探索Ultralytics如何与[Roboflow](https://roboflow.com/?ref=ultralytics)、ClearML、[Comet](https://bit.ly/yolov8-readme-comet)、Neural Magic以及[OpenVINO](https://docs.ultralytics.com/integrations/openvino/)合作,优化您的AI工作流程。
@@ -264,7 +264,7 @@ Ultralytics 提供了 YOLOv8 的交互式笔记本,涵盖训练、验证、跟 ##
贡献
-我们喜欢您的参与!没有社区的帮助,YOLOv5 和 YOLOv8 将无法实现。请参阅我们的[贡献指南](https://docs.ultralytics.com/help/contributing)以开始使用,并填写我们的[调查问卷](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey)向我们提供您的使用体验反馈。感谢所有贡献者的支持!🙏 +我们喜欢您的参与!没有社区的帮助,YOLOv5 和 YOLOv8 将无法实现。请参阅我们的[贡献指南](https://docs.ultralytics.com/help/contributing/)以开始使用,并填写我们的[调查问卷](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey)向我们提供您的使用体验反馈。感谢所有贡献者的支持!🙏 diff --git a/docs/README.md b/docs/README.md index 296b5d5a07..03285c41b4 100644 --- a/docs/README.md +++ b/docs/README.md @@ -3,7 +3,7 @@ # 📚 Ultralytics Docs -[Ultralytics](https://www.ultralytics.com/) Docs are the gateway to understanding and utilizing our cutting-edge machine learning tools. These documents are deployed to [https://docs.ultralytics.com](https://docs.ultralytics.com) for your convenience. +[Ultralytics](https://www.ultralytics.com/) Docs are the gateway to understanding and utilizing our cutting-edge machine learning tools. These documents are deployed to [https://docs.ultralytics.com](https://docs.ultralytics.com/) for your convenience. [![pages-build-deployment](https://github.com/ultralytics/docs/actions/workflows/pages/pages-build-deployment/badge.svg)](https://github.com/ultralytics/docs/actions/workflows/pages/pages-build-deployment) [![Check Broken links](https://github.com/ultralytics/docs/actions/workflows/links.yml/badge.svg)](https://github.com/ultralytics/docs/actions/workflows/links.yml) @@ -113,7 +113,7 @@ Choose a hosting provider and deployment method for your MkDocs documentation: ## 💡 Contribute -We cherish the community's input as it drives Ultralytics open-source initiatives. Dive into the [Contributing Guide](https://docs.ultralytics.com/help/contributing) and share your thoughts via our [Survey](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey). A heartfelt thank you 🙏 to each contributor! +We cherish the community's input as it drives Ultralytics open-source initiatives. Dive into the [Contributing Guide](https://docs.ultralytics.com/help/contributing/) and share your thoughts via our [Survey](https://www.ultralytics.com/survey?utm_source=github&utm_medium=social&utm_campaign=Survey). A heartfelt thank you 🙏 to each contributor! ![Ultralytics open-source contributors](https://github.com/ultralytics/docs/releases/download/0/ultralytics-open-source-contributors.avif) diff --git a/docs/coming_soon_template.md b/docs/coming_soon_template.md index 3f0840d453..5265d15cb8 100644 --- a/docs/coming_soon_template.md +++ b/docs/coming_soon_template.md @@ -27,7 +27,7 @@ Your feedback shapes our future releases. Share your thoughts and suggestions [h ## Thank You, Community! 🌍 -Your [contributions](https://docs.ultralytics.com/help/contributing) inspire our continuous [innovation](https://github.com/ultralytics/ultralytics). Stay tuned for the big reveal of what's next in AI and ML at Ultralytics! +Your [contributions](https://docs.ultralytics.com/help/contributing/) inspire our continuous [innovation](https://github.com/ultralytics/ultralytics). Stay tuned for the big reveal of what's next in AI and ML at Ultralytics! --- diff --git a/docs/en/datasets/index.md b/docs/en/datasets/index.md index f1e154ef7f..0a85ce15d0 100644 --- a/docs/en/datasets/index.md +++ b/docs/en/datasets/index.md @@ -123,7 +123,7 @@ Contributing a new dataset involves several steps to ensure that it aligns well 5. **Create a `data.yaml` File**: In your dataset's root directory, create a `data.yaml` file that describes the dataset, classes, and other necessary information. 6. **Optimize Images (Optional)**: If you want to reduce the size of the dataset for more efficient processing, you can optimize the images using the code below. This is not required, but recommended for smaller dataset sizes and faster download speeds. 7. **Zip Dataset**: Compress the entire dataset folder into a zip file. -8. **Document and PR**: Create a documentation page describing your dataset and how it fits into the existing framework. After that, submit a Pull Request (PR). Refer to [Ultralytics Contribution Guidelines](https://docs.ultralytics.com/help/contributing) for more details on how to submit a PR. +8. **Document and PR**: Create a documentation page describing your dataset and how it fits into the existing framework. After that, submit a Pull Request (PR). Refer to [Ultralytics Contribution Guidelines](https://docs.ultralytics.com/help/contributing/) for more details on how to submit a PR. ### Example Code to Optimize and Zip a Dataset @@ -175,7 +175,7 @@ Contributing a new dataset involves several steps: 5. **Create a `data.yaml` File**: Include dataset descriptions, classes, and other relevant information. 6. **Optimize Images (Optional)**: Reduce dataset size for efficiency. 7. **Zip Dataset**: Compress the dataset into a zip file. -8. **Document and PR**: Describe your dataset and submit a Pull Request following [Ultralytics Contribution Guidelines](https://docs.ultralytics.com/help/contributing). +8. **Document and PR**: Describe your dataset and submit a Pull Request following [Ultralytics Contribution Guidelines](https://docs.ultralytics.com/help/contributing/). Visit [Contribute New Datasets](#contribute-new-datasets) for a comprehensive guide. diff --git a/docs/en/datasets/segment/coco8-seg.md b/docs/en/datasets/segment/coco8-seg.md index fb505a5615..50a7f41af9 100644 --- a/docs/en/datasets/segment/coco8-seg.md +++ b/docs/en/datasets/segment/coco8-seg.md @@ -113,7 +113,7 @@ For a thorough explanation of available arguments and configuration options, you ### Why is the COCO8-Seg dataset important for model development and debugging? -The **COCO8-Seg dataset** is ideal for its manageability and diversity within a small size. It consists of only 8 images, providing a quick way to test and debug segmentation models or new detection approaches without the overhead of larger datasets. This makes it an efficient tool for sanity checks and pipeline error identification before committing to extensive training on large datasets. Learn more about dataset formats [here](https://docs.ultralytics.com/datasets/segment). +The **COCO8-Seg dataset** is ideal for its manageability and diversity within a small size. It consists of only 8 images, providing a quick way to test and debug segmentation models or new detection approaches without the overhead of larger datasets. This makes it an efficient tool for sanity checks and pipeline error identification before committing to extensive training on large datasets. Learn more about dataset formats [here](https://docs.ultralytics.com/datasets/segment/). ### Where can I find the YAML configuration file for the COCO8-Seg dataset? diff --git a/docs/en/datasets/segment/package-seg.md b/docs/en/datasets/segment/package-seg.md index ebdf4e6db2..db5f62be5f 100644 --- a/docs/en/datasets/segment/package-seg.md +++ b/docs/en/datasets/segment/package-seg.md @@ -136,7 +136,7 @@ This structure ensures a balanced dataset for thorough model training, validatio ### Why should I use Ultralytics YOLOv8 with the Package Segmentation Dataset? -Ultralytics YOLOv8 provides state-of-the-art accuracy and speed for real-time object detection and segmentation tasks. Using it with the Package Segmentation Dataset allows you to leverage YOLOv8's capabilities for precise package segmentation. This combination is especially beneficial for industries like logistics and warehouse automation, where accurate package identification is critical. For more information, check out our [page on YOLOv8 segmentation](https://docs.ultralytics.com/models/yolov8). +Ultralytics YOLOv8 provides state-of-the-art accuracy and speed for real-time object detection and segmentation tasks. Using it with the Package Segmentation Dataset allows you to leverage YOLOv8's capabilities for precise package segmentation. This combination is especially beneficial for industries like logistics and warehouse automation, where accurate package identification is critical. For more information, check out our [page on YOLOv8 segmentation](https://docs.ultralytics.com/models/yolov8/). ### How can I access and use the package-seg.yaml file for the Package Segmentation Dataset? diff --git a/docs/en/guides/object-blurring.md b/docs/en/guides/object-blurring.md index e0fe92041c..709ca1bd60 100644 --- a/docs/en/guides/object-blurring.md +++ b/docs/en/guides/object-blurring.md @@ -132,8 +132,8 @@ For more detailed applications, check the [advantages of object blurring section ### Can I use Ultralytics YOLOv8 to blur faces in a video for privacy reasons? -Yes, Ultralytics YOLOv8 can be configured to detect and blur faces in videos to protect privacy. By training or using a pre-trained model to specifically recognize faces, the detection results can be processed with OpenCV to apply a blur effect. Refer to our guide on [object detection with YOLOv8](https://docs.ultralytics.com/models/yolov8) and modify the code to target face detection. +Yes, Ultralytics YOLOv8 can be configured to detect and blur faces in videos to protect privacy. By training or using a pre-trained model to specifically recognize faces, the detection results can be processed with OpenCV to apply a blur effect. Refer to our guide on [object detection with YOLOv8](https://docs.ultralytics.com/models/yolov8/) and modify the code to target face detection. ### How does YOLOv8 compare to other object detection models like Faster R-CNN for object blurring? -Ultralytics YOLOv8 typically outperforms models like Faster R-CNN in terms of speed, making it more suitable for real-time applications. While both models offer accurate detection, YOLOv8's architecture is optimized for rapid inference, which is critical for tasks like real-time object blurring. Learn more about the technical differences and performance metrics in our [YOLOv8 documentation](https://docs.ultralytics.com/models/yolov8). +Ultralytics YOLOv8 typically outperforms models like Faster R-CNN in terms of speed, making it more suitable for real-time applications. While both models offer accurate detection, YOLOv8's architecture is optimized for rapid inference, which is critical for tasks like real-time object blurring. Learn more about the technical differences and performance metrics in our [YOLOv8 documentation](https://docs.ultralytics.com/models/yolov8/). diff --git a/docs/en/guides/object-counting.md b/docs/en/guides/object-counting.md index 68aae4fbae..749ffbe8fa 100644 --- a/docs/en/guides/object-counting.md +++ b/docs/en/guides/object-counting.md @@ -349,7 +349,7 @@ Ultralytics YOLOv8 provides several advantages over other object detection model 3. **Ease of Integration:** YOLOv8 offers seamless integration with various platforms and devices, including mobile and edge devices, which is crucial for modern AI applications. 4. **Flexibility:** Supports various tasks like object detection, segmentation, and tracking with configurable models to meet specific use-case requirements. -Check out Ultralytics [YOLOv8 Documentation](https://docs.ultralytics.com/models/yolov8) for a deeper dive into its features and performance comparisons. +Check out Ultralytics [YOLOv8 Documentation](https://docs.ultralytics.com/models/yolov8/) for a deeper dive into its features and performance comparisons. ### Can I use YOLOv8 for advanced applications like crowd analysis and traffic management? diff --git a/docs/en/guides/optimizing-openvino-latency-vs-throughput-modes.md b/docs/en/guides/optimizing-openvino-latency-vs-throughput-modes.md index 92c840f3a4..63fbb7a904 100644 --- a/docs/en/guides/optimizing-openvino-latency-vs-throughput-modes.md +++ b/docs/en/guides/optimizing-openvino-latency-vs-throughput-modes.md @@ -121,8 +121,8 @@ Using OpenVINO's high-level performance hints and multi-device modes can help st Yes, Ultralytics YOLO models are highly versatile and can be integrated with various AI frameworks. Options include: -- **TensorRT:** For NVIDIA GPU optimization, follow the [TensorRT integration guide](https://docs.ultralytics.com/integrations/tensorrt). -- **CoreML:** For Apple devices, refer to our [CoreML export instructions](https://docs.ultralytics.com/integrations/coreml). -- **TensorFlow.js:** For web and Node.js apps, see the [TF.js conversion guide](https://docs.ultralytics.com/integrations/tfjs). +- **TensorRT:** For NVIDIA GPU optimization, follow the [TensorRT integration guide](https://docs.ultralytics.com/integrations/tensorrt/). +- **CoreML:** For Apple devices, refer to our [CoreML export instructions](https://docs.ultralytics.com/integrations/coreml/). +- **TensorFlow.js:** For web and Node.js apps, see the [TF.js conversion guide](https://docs.ultralytics.com/integrations/tfjs/). -Explore more integrations on the [Ultralytics Integrations page](https://docs.ultralytics.com/integrations). +Explore more integrations on the [Ultralytics Integrations page](https://docs.ultralytics.com/integrations/). diff --git a/docs/en/guides/raspberry-pi.md b/docs/en/guides/raspberry-pi.md index 674bc847b3..25d45068a8 100644 --- a/docs/en/guides/raspberry-pi.md +++ b/docs/en/guides/raspberry-pi.md @@ -94,7 +94,7 @@ Here we will install Ultralytics package on the Raspberry Pi with optional depen ## Use NCNN on Raspberry Pi -Out of all the model export formats supported by Ultralytics, [NCNN](https://docs.ultralytics.com/integrations/ncnn) delivers the best inference performance when working with Raspberry Pi devices because NCNN is highly optimized for mobile/ embedded platforms (such as ARM architecture). Therefor our recommendation is to use NCNN with Raspberry Pi. +Out of all the model export formats supported by Ultralytics, [NCNN](https://docs.ultralytics.com/integrations/ncnn/) delivers the best inference performance when working with Raspberry Pi devices because NCNN is highly optimized for mobile/ embedded platforms (such as ARM architecture). Therefor our recommendation is to use NCNN with Raspberry Pi. ## Convert Model to NCNN and Run Inference @@ -132,7 +132,7 @@ The YOLOv8n model in PyTorch format is converted to NCNN to run inference with t !!! tip - For more details about supported export options, visit the [Ultralytics documentation page on deployment options](https://docs.ultralytics.com/guides/model-deployment-options). + For more details about supported export options, visit the [Ultralytics documentation page on deployment options](https://docs.ultralytics.com/guides/model-deployment-options/). ## Raspberry Pi 5 vs Raspberry Pi 4 YOLOv8 Benchmarks diff --git a/docs/en/guides/security-alarm-system.md b/docs/en/guides/security-alarm-system.md index 7e7051b63d..4b5cdeebd2 100644 --- a/docs/en/guides/security-alarm-system.md +++ b/docs/en/guides/security-alarm-system.md @@ -193,7 +193,7 @@ Running Ultralytics YOLOv8 on a standard setup typically requires around 5GB of ### What makes Ultralytics YOLOv8 different from other object detection models like Faster R-CNN or SSD? -Ultralytics YOLOv8 provides an edge over models like Faster R-CNN or SSD with its real-time detection capabilities and higher accuracy. Its unique architecture allows it to process images much faster without compromising on precision, making it ideal for time-sensitive applications like security alarm systems. For a comprehensive comparison of object detection models, you can explore our [guide](https://docs.ultralytics.com/models). +Ultralytics YOLOv8 provides an edge over models like Faster R-CNN or SSD with its real-time detection capabilities and higher accuracy. Its unique architecture allows it to process images much faster without compromising on precision, making it ideal for time-sensitive applications like security alarm systems. For a comprehensive comparison of object detection models, you can explore our [guide](https://docs.ultralytics.com/models/). ### How can I reduce the frequency of false positives in my security system using Ultralytics YOLOv8? diff --git a/docs/en/guides/streamlit-live-inference.md b/docs/en/guides/streamlit-live-inference.md index 3a89f123f4..84e1bce8ee 100644 --- a/docs/en/guides/streamlit-live-inference.md +++ b/docs/en/guides/streamlit-live-inference.md @@ -152,7 +152,7 @@ Real-time object detection using Streamlit and Ultralytics YOLOv8 can be applied - **Retail**: Customer counting, shelf management, and more. - **Wildlife and Agriculture**: Monitoring animals and crop conditions. -For more in-depth use cases and examples, explore [Ultralytics Solutions](https://docs.ultralytics.com/solutions). +For more in-depth use cases and examples, explore [Ultralytics Solutions](https://docs.ultralytics.com/solutions/). ### How does Ultralytics YOLOv8 compare to other object detection models like YOLOv5 and RCNNs? @@ -162,4 +162,4 @@ Ultralytics YOLOv8 provides several enhancements over prior models like YOLOv5 a - **Ease of Use**: Simplified interfaces and deployment. - **Resource Efficiency**: Optimized for better speed with minimal computational requirements. -For a comprehensive comparison, check [Ultralytics YOLOv8 Documentation](https://docs.ultralytics.com/models/yolov8) and related blog posts discussing model performance. +For a comprehensive comparison, check [Ultralytics YOLOv8 Documentation](https://docs.ultralytics.com/models/yolov8/) and related blog posts discussing model performance. diff --git a/docs/en/guides/triton-inference-server.md b/docs/en/guides/triton-inference-server.md index 1879bf78f3..05edf54f43 100644 --- a/docs/en/guides/triton-inference-server.md +++ b/docs/en/guides/triton-inference-server.md @@ -147,7 +147,7 @@ By following the above steps, you can deploy and run Ultralytics YOLOv8 models e ### How do I set up Ultralytics YOLOv8 with NVIDIA Triton Inference Server? -Setting up [Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) involves a few key steps: +Setting up [Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8/) with [NVIDIA Triton Inference Server](https://developer.nvidia.com/triton-inference-server) involves a few key steps: 1. **Export YOLOv8 to ONNX format**: @@ -258,7 +258,7 @@ For an in-depth guide on setting up and running Triton Server with YOLOv8, refer ### How does Ultralytics YOLOv8 compare to TensorFlow and PyTorch models for deployment? -[Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8) offers several unique advantages compared to TensorFlow and PyTorch models for deployment: +[Ultralytics YOLOv8](https://docs.ultralytics.com/models/yolov8/) offers several unique advantages compared to TensorFlow and PyTorch models for deployment: - **Real-time Performance**: Optimized for real-time object detection tasks, YOLOv8 provides state-of-the-art accuracy and speed, making it ideal for applications requiring live video analytics. - **Ease of Use**: YOLOv8 integrates seamlessly with Triton Inference Server and supports diverse export formats (ONNX, TensorRT, CoreML), making it flexible for various deployment scenarios. diff --git a/docs/en/help/contributing.md b/docs/en/help/contributing.md index 1d02ba23af..4078a6ce30 100644 --- a/docs/en/help/contributing.md +++ b/docs/en/help/contributing.md @@ -25,7 +25,7 @@ Welcome! We're thrilled that you're considering contributing to our [Ultralytics ## Code of Conduct -To ensure a welcoming and inclusive environment for everyone, all contributors must adhere to our [Code of Conduct](https://docs.ultralytics.com/help/code_of_conduct). Respect, kindness, and professionalism are at the heart of our community. +To ensure a welcoming and inclusive environment for everyone, all contributors must adhere to our [Code of Conduct](https://docs.ultralytics.com/help/code_of_conduct/). Respect, kindness, and professionalism are at the heart of our community. ## Contributing via Pull Requests @@ -45,7 +45,7 @@ We greatly appreciate contributions in the form of pull requests. To make the re ### CLA Signing -Before we can merge your pull request, you must sign our [Contributor License Agreement (CLA)](https://docs.ultralytics.com/help/CLA). This legal agreement ensures that your contributions are properly licensed, allowing the project to continue being distributed under the AGPL-3.0 license. +Before we can merge your pull request, you must sign our [Contributor License Agreement (CLA)](https://docs.ultralytics.com/help/CLA/). This legal agreement ensures that your contributions are properly licensed, allowing the project to continue being distributed under the AGPL-3.0 license. After submitting your pull request, the CLA bot will guide you through the signing process. To sign the CLA, simply add a comment in your PR stating: @@ -119,11 +119,11 @@ When adding new functions or classes, please include [Google-style docstrings](h ### GitHub Actions CI Tests -All pull requests must pass the GitHub Actions [Continuous Integration](https://docs.ultralytics.com/help/CI) (CI) tests before they can be merged. These tests include linting, unit tests, and other checks to ensure that your changes meet the project's quality standards. Review the CI output and address any issues that arise. +All pull requests must pass the GitHub Actions [Continuous Integration](https://docs.ultralytics.com/help/CI/) (CI) tests before they can be merged. These tests include linting, unit tests, and other checks to ensure that your changes meet the project's quality standards. Review the CI output and address any issues that arise. ## Reporting Bugs -We highly value bug reports as they help us maintain the quality of our projects. When reporting a bug, please provide a [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example)—a simple, clear code example that consistently reproduces the issue. This allows us to quickly identify and resolve the problem. +We highly value bug reports as they help us maintain the quality of our projects. When reporting a bug, please provide a [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example/)—a simple, clear code example that consistently reproduces the issue. This allows us to quickly identify and resolve the problem. ## License @@ -165,4 +165,4 @@ the project's quality standards. Review the CI output and fix any issues. For de ### How do I report a bug in Ultralytics YOLO repositories? -To report a bug, provide a clear and concise [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example) along with your bug report. This helps developers quickly identify and fix the issue. Ensure your example is minimal yet sufficient to replicate the problem. For more detailed steps on reporting bugs, refer to the [Reporting Bugs](#reporting-bugs) section. +To report a bug, provide a clear and concise [Minimum Reproducible Example](https://docs.ultralytics.com/help/minimum_reproducible_example/) along with your bug report. This helps developers quickly identify and fix the issue. Ensure your example is minimal yet sufficient to replicate the problem. For more detailed steps on reporting bugs, refer to the [Reporting Bugs](#reporting-bugs) section. diff --git a/docs/en/hub/datasets.md b/docs/en/hub/datasets.md index 40d76f5467..5e6f3c4c87 100644 --- a/docs/en/hub/datasets.md +++ b/docs/en/hub/datasets.md @@ -26,7 +26,7 @@ Once uploaded, datasets can be immediately utilized for model training. This int Before you upload a dataset to [Ultralytics HUB](https://www.ultralytics.com/hub), make sure to **place your dataset YAML file inside the dataset root directory** and that **your dataset YAML, directory and ZIP have the same name**, as shown in the example below, and then zip the dataset directory. -For example, if your dataset is called "coco8", as our [COCO8](https://docs.ultralytics.com/datasets/detect/coco8) example dataset, then you should have a `coco8.yaml` inside your `coco8/` directory, which will create a `coco8.zip` when zipped: +For example, if your dataset is called "coco8", as our [COCO8](https://docs.ultralytics.com/datasets/detect/coco8/) example dataset, then you should have a `coco8.yaml` inside your `coco8/` directory, which will create a `coco8.zip` when zipped: ```bash zip -r coco8.zip coco8 diff --git a/docs/en/hub/index.md b/docs/en/hub/index.md index bd3a4b62ab..77b98b61ce 100644 --- a/docs/en/hub/index.md +++ b/docs/en/hub/index.md @@ -120,7 +120,7 @@ Ultralytics HUB allows you to manage and organize your projects efficiently. You Ultralytics HUB offers seamless integrations with various platforms to enhance your machine learning workflows. Some key integrations include: - **Roboflow:** For dataset management and model training. Learn more on the [Integrations](integrations.md) page. -- **Google Colab:** Efficiently train models using Google Colab's cloud-based environment. Detailed steps are available in the [Google Colab](https://docs.ultralytics.com/integrations/google-colab) section. -- **Weights & Biases:** For enhanced experiment tracking and visualization. Explore the [Weights & Biases](https://docs.ultralytics.com/integrations/weights-biases) integration. +- **Google Colab:** Efficiently train models using Google Colab's cloud-based environment. Detailed steps are available in the [Google Colab](https://docs.ultralytics.com/integrations/google-colab/) section. +- **Weights & Biases:** For enhanced experiment tracking and visualization. Explore the [Weights & Biases](https://docs.ultralytics.com/integrations/weights-biases/) integration. For a complete list of integrations, refer to the [Integrations](integrations.md) page. diff --git a/docs/en/hub/inference-api.md b/docs/en/hub/inference-api.md index 4146599f88..cd98347efe 100644 --- a/docs/en/hub/inference-api.md +++ b/docs/en/hub/inference-api.md @@ -39,7 +39,7 @@ To use the [Ultralytics HUB](https://www.ultralytics.com/hub) Dedicated Inferenc !!! tip - Choose the region with the lowest latency for the best performance as described in the [documentation](https://docs.ultralytics.com/reference/hub/google/__init__). + Choose the region with the lowest latency for the best performance as described in the [documentation](https://docs.ultralytics.com/reference/hub/google/__init__/). To shut down the dedicated endpoint, click on the **Stop Endpoint** button. diff --git a/docs/en/hub/integrations.md b/docs/en/hub/integrations.md index 6885f07639..a86eeed9f1 100644 --- a/docs/en/hub/integrations.md +++ b/docs/en/hub/integrations.md @@ -124,4 +124,4 @@ Your feedback shapes our future releases. Share your thoughts and suggestions [h ## Thank You, Community! 🌍 -Your [contributions](https://docs.ultralytics.com/help/contributing) inspire our continuous [innovation](https://github.com/ultralytics/ultralytics). Stay tuned for the big reveal of what's next in AI and ML at Ultralytics! +Your [contributions](https://docs.ultralytics.com/help/contributing/) inspire our continuous [innovation](https://github.com/ultralytics/ultralytics). Stay tuned for the big reveal of what's next in AI and ML at Ultralytics! diff --git a/docs/en/hub/models.md b/docs/en/hub/models.md index 90dfc792bb..eae6612608 100644 --- a/docs/en/hub/models.md +++ b/docs/en/hub/models.md @@ -66,9 +66,9 @@ In this step, you have to choose the project in which you want to create your mo !!! info - You can read more about the available [YOLOv8](https://docs.ultralytics.com/models/yolov8) (and [YOLOv5](https://docs.ultralytics.com/models/yolov5)) architectures in our documentation. + You can read more about the available [YOLOv8](https://docs.ultralytics.com/models/yolov8/) (and [YOLOv5](https://docs.ultralytics.com/models/yolov5/)) architectures in our documentation. -By default, your model will use a pre-trained model (trained on the [COCO](https://docs.ultralytics.com/datasets/detect/coco) dataset) to reduce training time. You can change this behavior and tweak your model's configuration by opening the **Advanced Model Configuration** accordion. +By default, your model will use a pre-trained model (trained on the [COCO](https://docs.ultralytics.com/datasets/detect/coco/) dataset) to reduce training time. You can change this behavior and tweak your model's configuration by opening the **Advanced Model Configuration** accordion. ![Ultralytics HUB screenshot of the Train Model dialog with an arrow pointing to the Advanced Model Configuration accordion](https://github.com/ultralytics/docs/releases/download/0/ultralytics-hub-train-model-dialog-2.avif) diff --git a/docs/en/hub/quickstart.md b/docs/en/hub/quickstart.md index e73105d561..f951e97d53 100644 --- a/docs/en/hub/quickstart.md +++ b/docs/en/hub/quickstart.md @@ -54,7 +54,7 @@ You can upload a dataset directly from the [Home](https://hub.ultralytics.com/ho ![Ultralytics HUB screenshot of the Home page with an arrow pointing to the Upload Dataset card](https://github.com/ultralytics/docs/releases/download/0/ultralytics-hub-upload-dataset-card.avif) -Read more about [datasets](https://docs.ultralytics.com/hub/datasets). +Read more about [datasets](https://docs.ultralytics.com/hub/datasets/). ### Create Project @@ -62,7 +62,7 @@ You can create a project directly from the [Home](https://hub.ultralytics.com/ho ![Ultralytics HUB screenshot of the Home page with an arrow pointing to the Create Project card](https://github.com/ultralytics/docs/releases/download/0/hub-create-project-card.avif) -Read more about [projects](https://docs.ultralytics.com/hub/projects). +Read more about [projects](https://docs.ultralytics.com/hub/projects/). ### Train Model @@ -70,7 +70,7 @@ You can train a model directly from the [Home](https://hub.ultralytics.com/home) ![Ultralytics HUB screenshot of the Home page with an arrow pointing to the Train Model card](https://github.com/ultralytics/docs/releases/download/0/ultralytics-hub-train-model-card.avif) -Read more about [models](https://docs.ultralytics.com/hub/models). +Read more about [models](https://docs.ultralytics.com/hub/models/). ## Feedback diff --git a/docs/en/integrations/edge-tpu.md b/docs/en/integrations/edge-tpu.md index d1caf90f5c..df63049b06 100644 --- a/docs/en/integrations/edge-tpu.md +++ b/docs/en/integrations/edge-tpu.md @@ -113,7 +113,7 @@ However, for in-depth instructions on deploying your TFLite Edge TPU models, tak In this guide, we've learned how to export Ultralytics YOLOv8 models to TFLite Edge TPU format. By following the steps mentioned above, you can increase the speed and power of your computer vision applications. -For further details on usage, visit the [Edge TPU official website](https://cloud.google.com/edge-tpu). +For further details on usage, visit the [Edge TPU official website](https://cloud.google.com/tpu). Also, for more information on other Ultralytics YOLOv8 integrations, please visit our [integration guide page](index.md). There, you'll discover valuable resources and insights. diff --git a/docs/en/modes/export.md b/docs/en/modes/export.md index 443343947a..a29c4c70d9 100644 --- a/docs/en/modes/export.md +++ b/docs/en/modes/export.md @@ -46,7 +46,7 @@ Here are some of the standout functionalities: ## Usage Examples -Export a YOLOv8n model to a different format like ONNX or TensorRT. See Arguments section below for a full list of export arguments. +Export a YOLOv8n model to a different format like ONNX or TensorRT. See the Arguments section below for a full list of export arguments. !!! example @@ -112,7 +112,7 @@ Exporting a YOLOv8 model to ONNX format is straightforward with Ultralytics. It yolo export model=path/to/best.pt format=onnx # export custom trained model ``` -For more details on the process, including advanced options like handling different input sizes, refer to the [ONNX](../integrations/onnx.md) section. +For more details on the process, including advanced options like handling different input sizes, refer to the [ONNX section](../integrations/onnx.md). ### What are the benefits of using TensorRT for model export? @@ -122,7 +122,7 @@ Using TensorRT for model export offers significant performance improvements. YOL - **Speed:** Achieve faster inference through advanced optimizations. - **Compatibility:** Integrate smoothly with NVIDIA hardware. -To learn more about integrating TensorRT, see the [TensorRT](../integrations/tensorrt.md) integration guide. +To learn more about integrating TensorRT, see the [TensorRT integration guide](../integrations/tensorrt.md). ### How do I enable INT8 quantization when exporting my YOLOv8 model? @@ -145,7 +145,7 @@ INT8 quantization is an excellent way to compress the model and speed up inferen yolo export model=yolov8n.pt format=onnx int8=True # export model with INT8 quantization ``` -INT8 quantization can be applied to various formats, such as TensorRT and CoreML. More details can be found in the [Export](../modes/export.md) section. +INT8 quantization can be applied to various formats, such as TensorRT and CoreML. More details can be found in the [Export section](../modes/export.md). ### Why is dynamic input size important when exporting models? @@ -182,4 +182,4 @@ Understanding and configuring export arguments is crucial for optimizing model p - **`optimize:`** Applies specific optimizations for mobile or constrained environments. - **`int8:`** Enables INT8 quantization, highly beneficial for edge deployments. -For a detailed list and explanations of all the export arguments, visit the [Export Arguments](#arguments) section. +For a detailed list and explanations of all the export arguments, visit the [Export Arguments section](#arguments). diff --git a/docs/en/modes/index.md b/docs/en/modes/index.md index 270ce4a8ee..6797ac19df 100644 --- a/docs/en/modes/index.md +++ b/docs/en/modes/index.md @@ -68,7 +68,7 @@ Track mode is used for tracking objects in real-time using a YOLOv8 model. In th ## [Benchmark](benchmark.md) -Benchmark mode is used to profile the speed and accuracy of various export formats for YOLOv8. The benchmarks provide information on the size of the exported format, its `mAP50-95` metrics (for object detection, segmentation and pose) or `accuracy_top5` metrics (for classification), and the inference time in milliseconds per image across various export formats like ONNX, OpenVINO, TensorRT and others. This information can help users choose the optimal export format for their specific use case based on their requirements for speed and accuracy. +Benchmark mode is used to profile the speed and accuracy of various export formats for YOLOv8. The benchmarks provide information on the size of the exported format, its `mAP50-95` metrics (for object detection, segmentation, and pose) or `accuracy_top5` metrics (for classification), and the inference time in milliseconds per image across various formats like ONNX, OpenVINO, TensorRT, and others. This information can help users choose the optimal export format for their specific use case based on their requirements for speed and accuracy. [Benchmark Examples](benchmark.md){ .md-button } diff --git a/docs/en/modes/train.md b/docs/en/modes/train.md index 7ac428d66e..03e3be223d 100644 --- a/docs/en/modes/train.md +++ b/docs/en/modes/train.md @@ -123,14 +123,14 @@ To enable training on Apple M1 and M2 chips, you should specify 'mps' as your de # Load a model model = YOLO("yolov8n.pt") # load a pretrained model (recommended for training) - # Train the model with 2 GPUs + # Train the model with MPS results = model.train(data="coco8.yaml", epochs=100, imgsz=640, device="mps") ``` === "CLI" ```bash - # Start training from a pretrained *.pt model using GPUs 0 and 1 + # Start training from a pretrained *.pt model using MPS yolo detect train data=coco8.yaml model=yolov8n.pt epochs=100 imgsz=640 device=mps ``` @@ -169,7 +169,7 @@ Below is an example of how to resume an interrupted training using Python and vi By setting `resume=True`, the `train` function will continue training from where it left off, using the state stored in the 'path/to/last.pt' file. If the `resume` argument is omitted or set to `False`, the `train` function will start a new training session. -Remember that checkpoints are saved at the end of every epoch by default, or at fixed interval using the `save_period` argument, so you must complete at least 1 epoch to resume a training run. +Remember that checkpoints are saved at the end of every epoch by default, or at fixed intervals using the `save_period` argument, so you must complete at least 1 epoch to resume a training run. ## Train Settings diff --git a/docs/en/modes/val.md b/docs/en/modes/val.md index 1cbd900d96..85b15f51e5 100644 --- a/docs/en/modes/val.md +++ b/docs/en/modes/val.md @@ -47,7 +47,7 @@ These are the notable functionalities offered by YOLOv8's Val mode: ## Usage Examples -Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. See Arguments section below for a full list of export arguments. +Validate trained YOLOv8n model accuracy on the COCO8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. See Arguments section below for a full list of export arguments. !!! example @@ -165,7 +165,7 @@ These benefits ensure that your models are evaluated thoroughly and can be optim ### Can I validate my YOLOv8 model using a custom dataset? -Yes, you can validate your YOLOv8 model using a custom dataset. Specify the `data` argument with the path to your dataset configuration file. This file should include paths to the validation data, class names, and other relevant details. +Yes, you can validate your YOLOv8 model using a [custom dataset](https://docs.ultralytics.com/datasets/). Specify the `data` argument with the path to your dataset configuration file. This file should include paths to the validation data, class names, and other relevant details. Example in Python: diff --git a/docs/en/tasks/classify.md b/docs/en/tasks/classify.md index 07a9cc4ecb..7df4b2f7c9 100644 --- a/docs/en/tasks/classify.md +++ b/docs/en/tasks/classify.md @@ -84,7 +84,7 @@ YOLO classification dataset format can be found in detail in the [Dataset Guide] ## Val -Validate trained YOLOv8n-cls model accuracy on the MNIST160 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n-cls model accuracy on the MNIST160 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/tasks/detect.md b/docs/en/tasks/detect.md index d3542f184d..a46956a849 100644 --- a/docs/en/tasks/detect.md +++ b/docs/en/tasks/detect.md @@ -83,7 +83,7 @@ YOLO detection dataset format can be found in detail in the [Dataset Guide](../d ## Val -Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n model accuracy on the COCO8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/tasks/obb.md b/docs/en/tasks/obb.md index a603ab1a84..2fc2b9da21 100644 --- a/docs/en/tasks/obb.md +++ b/docs/en/tasks/obb.md @@ -104,8 +104,7 @@ OBB dataset format can be found in detail in the [Dataset Guide](../datasets/obb ## Val -Validate trained YOLOv8n-obb model accuracy on the DOTA8 dataset. No argument need to passed as the `model` -retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n-obb model accuracy on the DOTA8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/tasks/pose.md b/docs/en/tasks/pose.md index 81384cc951..ead809d418 100644 --- a/docs/en/tasks/pose.md +++ b/docs/en/tasks/pose.md @@ -117,8 +117,7 @@ YOLO pose dataset format can be found in detail in the [Dataset Guide](../datase ## Val -Validate trained YOLOv8n-pose model accuracy on the COCO128-pose dataset. No argument need to passed as the `model` -retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n-pose model accuracy on the COCO128-pose dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/tasks/segment.md b/docs/en/tasks/segment.md index 56a3b3ba24..e1c43eea0c 100644 --- a/docs/en/tasks/segment.md +++ b/docs/en/tasks/segment.md @@ -84,8 +84,7 @@ YOLO segmentation dataset format can be found in detail in the [Dataset Guide](. ## Val -Validate trained YOLOv8n-seg model accuracy on the COCO128-seg dataset. No argument need to passed as the `model` -retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n-seg model accuracy on the COCO128-seg dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/usage/cli.md b/docs/en/usage/cli.md index f22e3eb4cf..3e6c4b37a5 100644 --- a/docs/en/usage/cli.md +++ b/docs/en/usage/cli.md @@ -109,7 +109,7 @@ Train YOLOv8n on the COCO8 dataset for 100 epochs at image size 640. For a full ## Val -Validate trained YOLOv8n model accuracy on the COCO8 dataset. No argument need to passed as the `model` retains its training `data` and arguments as model attributes. +Validate trained YOLOv8n model accuracy on the COCO8 dataset. No arguments are needed as the `model` retains its training `data` and arguments as model attributes. !!! example diff --git a/docs/en/yolov5/tutorials/train_custom_data.md b/docs/en/yolov5/tutorials/train_custom_data.md index dbe62fc25c..6119bc3a7c 100644 --- a/docs/en/yolov5/tutorials/train_custom_data.md +++ b/docs/en/yolov5/tutorials/train_custom_data.md @@ -42,7 +42,7 @@ YOLOv5 models must be trained on labelled data in order to learn classes of obje Your model will learn by example. Training on images similar to the ones it will see in the wild is of the utmost importance. Ideally, you will collect a wide variety of images from the same configuration (camera, angle, lighting, etc.) as you will ultimately deploy your project. -If this is not possible, you can start from [a public dataset](https://universe.roboflow.com/?ref=ultralytics) to train your initial model and then [sample images from the wild during inference](https://blog.roboflow.com/computer-vision-active-learning-tips/?ref=ultralytics) to improve your dataset and model iteratively. +If this is not possible, you can start from [a public dataset](https://universe.roboflow.com/?ref=ultralytics) to train your initial model and then [sample images from the wild during inference](https://blog.roboflow.com/what-is-active-learning/?ref=ultralytics) to improve your dataset and model iteratively. ### 1.2 Create Labels diff --git a/examples/README.md b/examples/README.md index 3e3a942365..931bdc634c 100644 --- a/examples/README.md +++ b/examples/README.md @@ -32,6 +32,6 @@ We greatly appreciate contributions from the community, including examples, appl - Avoids adding large files or dependencies unless they are absolutely necessary for the example. - Contributors should be willing to provide support for their examples and address related issues. -For more detailed information and guidance on contributing, please visit our [contribution documentation](https://docs.ultralytics.com/help/contributing). +For more detailed information and guidance on contributing, please visit our [contribution documentation](https://docs.ultralytics.com/help/contributing/). If you encounter any questions or concerns regarding these guidelines, feel free to open a PR or an issue in the repository, and we will assist you in the contribution process. diff --git a/pyproject.toml b/pyproject.toml index 03e557e941..8c915c0d3c 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -26,11 +26,11 @@ build-backend = "setuptools.build_meta" [project] name = "ultralytics" dynamic = ["version"] -description = "Ultralytics YOLOv8 for SOTA object detection, multi-object tracking, instance segmentation, pose estimation and image classification." +description = "Ultralytics YOLO for SOTA object detection, multi-object tracking, instance segmentation, pose estimation and image classification." readme = "README.md" requires-python = ">=3.8" license = { "text" = "AGPL-3.0" } -keywords = ["machine-learning", "deep-learning", "computer-vision", "ML", "DL", "AI", "YOLO", "YOLOv3", "YOLOv5", "YOLOv8", "HUB", "Ultralytics"] +keywords = ["machine-learning", "deep-learning", "computer-vision", "ML", "DL", "AI", "YOLO", "YOLOv3", "YOLOv5", "YOLOv8", "YOLOv9", "YOLOv10", "HUB", "Ultralytics"] authors = [ { name = "Glenn Jocher" }, { name = "Ayush Chaurasia" }, diff --git a/ultralytics/cfg/models/README.md b/ultralytics/cfg/models/README.md index 274a594eb8..bcaf8deda4 100644 --- a/ultralytics/cfg/models/README.md +++ b/ultralytics/cfg/models/README.md @@ -4,11 +4,11 @@ Welcome to the [Ultralytics](https://www.ultralytics.com/) Models directory! Her These model configurations cover a wide range of scenarios, from simple object detection to more complex tasks like instance segmentation and object tracking. They are also designed to run efficiently on a variety of hardware platforms, from CPUs to GPUs. Whether you are a seasoned machine learning practitioner or just getting started with YOLO, this directory provides a great starting point for your custom model development needs. -To get started, simply browse through the models in this directory and find one that best suits your needs. Once you've selected a model, you can use the provided `*.yaml` file to train and deploy your custom YOLO model with ease. See full details at the Ultralytics [Docs](https://docs.ultralytics.com/models), and if you need help or have any questions, feel free to reach out to the Ultralytics team for support. So, don't wait, start creating your custom YOLO model now! +To get started, simply browse through the models in this directory and find one that best suits your needs. Once you've selected a model, you can use the provided `*.yaml` file to train and deploy your custom YOLO model with ease. See full details at the Ultralytics [Docs](https://docs.ultralytics.com/models/), and if you need help or have any questions, feel free to reach out to the Ultralytics team for support. So, don't wait, start creating your custom YOLO model now! ### Usage -Model `*.yaml` files may be used directly in the [Command Line Interface (CLI)](https://docs.ultralytics.com/usage/cli) with a `yolo` command: +Model `*.yaml` files may be used directly in the [Command Line Interface (CLI)](https://docs.ultralytics.com/usage/cli/) with a `yolo` command: ```bash # Train a YOLOv8n model using the coco8 dataset for 100 epochs @@ -35,7 +35,7 @@ model.train(data="coco8.yaml", epochs=100) ## Pre-trained Model Architectures -Ultralytics supports many model architectures. Visit [Ultralytics Models](https://docs.ultralytics.com/models) to view detailed information and usage. Any of these models can be used by loading their configurations or pretrained checkpoints if available. +Ultralytics supports many model architectures. Visit [Ultralytics Models](https://docs.ultralytics.com/models/) to view detailed information and usage. Any of these models can be used by loading their configurations or pretrained checkpoints if available. ## Contribute New Models @@ -43,6 +43,6 @@ Have you trained a new YOLO variant or achieved state-of-the-art performance wit By contributing to this section, you're helping us offer a wider array of model choices and configurations to the community. It's a fantastic way to share your knowledge and expertise while making the Ultralytics YOLO ecosystem even more versatile. -To get started, please consult our [Contributing Guide](https://docs.ultralytics.com/help/contributing) for step-by-step instructions on how to submit a Pull Request (PR) 🛠️. Your contributions are eagerly awaited! +To get started, please consult our [Contributing Guide](https://docs.ultralytics.com/help/contributing/) for step-by-step instructions on how to submit a Pull Request (PR) 🛠️. Your contributions are eagerly awaited! Let's join hands to extend the range and capabilities of the Ultralytics YOLO models 🙏! diff --git a/ultralytics/trackers/README.md b/ultralytics/trackers/README.md index 1675519c6b..d7bc855814 100644 --- a/ultralytics/trackers/README.md +++ b/ultralytics/trackers/README.md @@ -309,6 +309,6 @@ Are you proficient in multi-object tracking and have successfully implemented or By contributing to this section, you help expand the scope of tracking solutions available within the Ultralytics YOLO framework, adding another layer of functionality and utility for the community. -To initiate your contribution, please refer to our [Contributing Guide](https://docs.ultralytics.com/help/contributing) for comprehensive instructions on submitting a Pull Request (PR) 🛠️. We are excited to see what you bring to the table! +To initiate your contribution, please refer to our [Contributing Guide](https://docs.ultralytics.com/help/contributing/) for comprehensive instructions on submitting a Pull Request (PR) 🛠️. We are excited to see what you bring to the table! Together, let's enhance the tracking capabilities of the Ultralytics YOLO ecosystem 🙏!