You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 

3.2 KiB

YOLOv8 on Jetson

Only test on Jetson-NX 4GB

ENVS:

  • Jetpack 4.6.3
  • CUDA-10.2
  • CUDNN-8.2.1
  • TensorRT-8.2.1
  • DeepStream-6.0.1
  • OpenCV-4.1.1
  • CMake-3.10.2

If you have other environment-related issues, please discuss in issue.

End2End Detection

1. Export Detection End2End ONNX

yolov8s.pt is your trained pytorch model, or the official pre-trained model.

Do not use any model other than pytorch model. Do not use build.py to export engine if you don't know how to install pytorch and other environments on jetson.

!!! Please use the PC to execute the following script !!!

# Export yolov8s.pt to yolov8s.onnx
python3 export-det.py --weights yolov8s.pt --sim

!!! Please use the Jetson to execute the following script !!!

# Using trtexec tools for export engine
/usr/src/tensorrt/bin/trtexec \
--onnx=yolov8s.onnx \
--saveEngine=yolov8s.engine

After executing the above command, you will get an engine named yolov8s.engine .

2. Inference with c++

It is highly recommended to use C++ inference on Jetson. Here is a demo: csrc/jetson/detect .

Build:

Please modify CLASS_NAMES and COLORS in main.cpp for yourself.

And build:

export root=${PWD}
cd src/jetson/detect
mkdir build
cmake ..
make
mv yolov8 ${root}
cd ${root}

Usage:

# infer image
./yolov8 yolov8s.engine data/bus.jpg
# infer images
./yolov8 yolov8s.engine data
# infer video
./yolov8 yolov8s.engine data/test.mp4 # the video path

Speedup Segmention

1. Export Segmention Speedup ONNX

yolov8s-seg.pt is your trained pytorch model, or the official pre-trained model.

Do not use any model other than pytorch model. Do not use build.py to export engine if you don't know how to install pytorch and other environments on jetson.

!!! Please use the PC to execute the following script !!!

# Export yolov8s-seg.pt to yolov8s-seg.onnx
python3 export-seg.py --weights yolov8s-seg.pt --sim

!!! Please use the Jetson to execute the following script !!!

# Using trtexec tools for export engine
/usr/src/tensorrt/bin/trtexec \
--onnx=yolov8s-seg.onnx \
--saveEngine=yolov8s-seg.engine

After executing the above command, you will get an engine named yolov8s-seg.engine .

2. Inference with c++

It is highly recommended to use C++ inference on Jetson. Here is a demo: csrc/jetson/segment .

Build:

Please modify CLASS_NAMES and COLORS and postprocess parameters in main.cpp for yourself.

int topk = 100;
int seg_h = 160; // yolov8 model proto height
int seg_w = 160; // yolov8 model proto width
int seg_channels = 32; // yolov8 model proto channels
float score_thres = 0.25f;
float iou_thres = 0.65f;

And build:

export root=${PWD}
cd src/jetson/segment
mkdir build
cmake ..
make
mv yolov8-seg ${root}
cd ${root}

Usage:

# infer image
./yolov8-seg yolov8s-seg.engine data/bus.jpg
# infer images
./yolov8-seg yolov8s-seg.engine data
# infer video
./yolov8-seg yolov8s-seg.engine data/test.mp4 # the video path