You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
139 lines
6.8 KiB
139 lines
6.8 KiB
2 years ago
|
---
|
||
|
comments: true
|
||
2 years ago
|
description: Understand the YOLO pose dataset format and learn to use Ultralytics datasets to train your pose estimation models effectively.
|
||
|
keywords: Ultralytics, YOLO, pose estimation, datasets, training, YAML, keypoints, COCO-Pose, COCO8-Pose, data conversion
|
||
2 years ago
|
---
|
||
|
|
||
|
# Pose Estimation Datasets Overview
|
||
|
|
||
|
## Supported Dataset Formats
|
||
|
|
||
|
### Ultralytics YOLO format
|
||
|
|
||
2 years ago
|
The dataset label format used for training YOLO pose models is as follows:
|
||
2 years ago
|
|
||
|
1. One text file per image: Each image in the dataset has a corresponding text file with the same name as the image file and the ".txt" extension.
|
||
|
2. One row per object: Each row in the text file corresponds to one object instance in the image.
|
||
|
3. Object information per row: Each row contains the following information about the object instance:
|
||
2 years ago
|
- Object class index: An integer representing the class of the object (e.g., 0 for person, 1 for car, etc.).
|
||
|
- Object center coordinates: The x and y coordinates of the center of the object, normalized to be between 0 and 1.
|
||
|
- Object width and height: The width and height of the object, normalized to be between 0 and 1.
|
||
|
- Object keypoint coordinates: The keypoints of the object, normalized to be between 0 and 1.
|
||
2 years ago
|
|
||
|
Here is an example of the label format for pose estimation task:
|
||
|
|
||
|
Format with Dim = 2
|
||
|
|
||
|
```
|
||
2 years ago
|
<class-index> <x> <y> <width> <height> <px1> <py1> <px2> <py2> ... <pxn> <pyn>
|
||
2 years ago
|
```
|
||
2 years ago
|
|
||
2 years ago
|
Format with Dim = 3
|
||
|
|
||
|
```
|
||
|
<class-index> <x> <y> <width> <height> <px1> <py1> <p1-visibility> <px2> <py2> <p2-visibility> <pxn> <pyn> <p2-visibility>
|
||
|
```
|
||
|
|
||
1 year ago
|
In this format, `<class-index>` is the index of the class for the object,`<x> <y> <width> <height>` are coordinates of bounding box, and `<px1> <py1> <px2> <py2> ... <pxn> <pyn>` are the pixel coordinates of the keypoints. The coordinates are separated by spaces.
|
||
2 years ago
|
|
||
2 years ago
|
### Dataset YAML format
|
||
2 years ago
|
|
||
|
The Ultralytics framework uses a YAML file format to define the dataset and model configuration for training Detection Models. Here is an example of the YAML format used for defining a detection dataset:
|
||
|
|
||
|
```yaml
|
||
2 years ago
|
# Train/val/test sets as 1) dir: path/to/imgs, 2) file: path/to/imgs.txt, or 3) list: [path/to/imgs1, path/to/imgs2, ..]
|
||
|
path: ../datasets/coco8-pose # dataset root dir
|
||
|
train: images/train # train images (relative to 'path') 4 images
|
||
|
val: images/val # val images (relative to 'path') 4 images
|
||
|
test: # test images (optional)
|
||
2 years ago
|
|
||
|
# Keypoints
|
||
2 years ago
|
kpt_shape: [17, 3] # number of keypoints, number of dims (2 for x,y or 3 for x,y,visible)
|
||
|
flip_idx: [0, 2, 1, 4, 3, 6, 5, 8, 7, 10, 9, 12, 11, 14, 13, 16, 15]
|
||
2 years ago
|
|
||
2 years ago
|
# Classes dictionary
|
||
2 years ago
|
names:
|
||
|
0: person
|
||
|
```
|
||
|
|
||
2 years ago
|
The `train` and `val` fields specify the paths to the directories containing the training and validation images, respectively.
|
||
2 years ago
|
|
||
2 years ago
|
`names` is a dictionary of class names. The order of the names should match the order of the object class indices in the YOLO dataset files.
|
||
2 years ago
|
|
||
1 year ago
|
(Optional) if the points are symmetric then need flip_idx, like left-right side of human or face. For example if we assume five keypoints of facial landmark: [left eye, right eye, nose, left mouth, right mouth], and the original index is [0, 1, 2, 3, 4], then flip_idx is [1, 0, 2, 4, 3] (just exchange the left-right index, i.e 0-1 and 3-4, and do not modify others like nose in this example).
|
||
2 years ago
|
|
||
|
## Usage
|
||
2 years ago
|
|
||
2 years ago
|
!!! example ""
|
||
|
|
||
|
=== "Python"
|
||
2 years ago
|
|
||
2 years ago
|
```python
|
||
|
from ultralytics import YOLO
|
||
2 years ago
|
|
||
2 years ago
|
# Load a model
|
||
|
model = YOLO('yolov8n-pose.pt') # load a pretrained model (recommended for training)
|
||
|
|
||
|
# Train the model
|
||
2 years ago
|
results = model.train(data='coco128-pose.yaml', epochs=100, imgsz=640)
|
||
2 years ago
|
```
|
||
|
=== "CLI"
|
||
2 years ago
|
|
||
2 years ago
|
```bash
|
||
|
# Start training from a pretrained *.pt model
|
||
|
yolo detect train data=coco128-pose.yaml model=yolov8n-pose.pt epochs=100 imgsz=640
|
||
|
```
|
||
|
|
||
|
## Supported Datasets
|
||
2 years ago
|
|
||
2 years ago
|
This section outlines the datasets that are compatible with Ultralytics YOLO format and can be used for training pose estimation models:
|
||
2 years ago
|
|
||
2 years ago
|
### COCO-Pose
|
||
2 years ago
|
|
||
2 years ago
|
- **Description**: COCO-Pose is a large-scale object detection, segmentation, and pose estimation dataset. It is a subset of the popular COCO dataset and focuses on human pose estimation. COCO-Pose includes multiple keypoints for each human instance.
|
||
|
- **Label Format**: Same as Ultralytics YOLO format as described above, with keypoints for human poses.
|
||
|
- **Number of Classes**: 1 (Human).
|
||
|
- **Keypoints**: 17 keypoints including nose, eyes, ears, shoulders, elbows, wrists, hips, knees, and ankles.
|
||
|
- **Usage**: Suitable for training human pose estimation models.
|
||
|
- **Additional Notes**: The dataset is rich and diverse, containing over 200k labeled images.
|
||
|
- [Read more about COCO-Pose](./coco.md)
|
||
|
|
||
|
### COCO8-Pose
|
||
|
|
||
|
- **Description**: [Ultralytics](https://ultralytics.com) COCO8-Pose is a small, but versatile pose detection dataset composed of the first 8 images of the COCO train 2017 set, 4 for training and 4 for validation.
|
||
|
- **Label Format**: Same as Ultralytics YOLO format as described above, with keypoints for human poses.
|
||
|
- **Number of Classes**: 1 (Human).
|
||
|
- **Keypoints**: 17 keypoints including nose, eyes, ears, shoulders, elbows, wrists, hips, knees, and ankles.
|
||
|
- **Usage**: Suitable for testing and debugging object detection models, or for experimenting with new detection approaches.
|
||
|
- **Additional Notes**: COCO8-Pose is ideal for sanity checks and CI checks.
|
||
|
- [Read more about COCO8-Pose](./coco8-pose.md)
|
||
|
|
||
1 year ago
|
### Tiger-Pose
|
||
|
|
||
|
- **Description**: [Ultralytics](https://ultralytics.com) This animal pose dataset comprises 263 images sourced from a [YouTube Video](https://www.youtube.com/watch?v=MIBAT6BGE6U&pp=ygUbVGlnZXIgd2Fsa2luZyByZWZlcmVuY2UubXA0), with 210 images allocated for training and 53 for validation.
|
||
|
- **Label Format**: Same as Ultralytics YOLO format as described above, with 12 keypoints for animal pose and no visible dimension.
|
||
|
- **Number of Classes**: 1 (Tiger).
|
||
|
- **Keypoints**: 12 keypoints.
|
||
|
- **Usage**: Great for animal pose or any other pose that is not human-based.
|
||
|
- [Read more about Tiger-Pose](./tiger-pose.md)
|
||
|
|
||
2 years ago
|
### Adding your own dataset
|
||
|
|
||
|
If you have your own dataset and would like to use it for training pose estimation models with Ultralytics YOLO format, ensure that it follows the format specified above under "Ultralytics YOLO format". Convert your annotations to the required format and specify the paths, number of classes, and class names in the YAML configuration file.
|
||
|
|
||
|
### Conversion Tool
|
||
|
|
||
|
Ultralytics provides a convenient conversion tool to convert labels from the popular COCO dataset format to YOLO format:
|
||
2 years ago
|
|
||
2 years ago
|
!!! example ""
|
||
2 years ago
|
|
||
2 years ago
|
=== "Python"
|
||
|
|
||
|
```python
|
||
|
from ultralytics.data.converter import convert_coco
|
||
1 year ago
|
|
||
2 years ago
|
convert_coco(labels_dir='path/to/coco/annotations/', use_keypoints=True)
|
||
|
```
|
||
2 years ago
|
|
||
2 years ago
|
This conversion tool can be used to convert the COCO dataset or any dataset in the COCO format to the Ultralytics YOLO format. The `use_keypoints` parameter specifies whether to include keypoints (for pose estimation) in the converted labels.
|