The simplest way of simply using YOLOv8 directly in a Python environment. !!! example "Train" === "From pretrained(recommanded)" ```python from ultralytics import YOLO model = YOLO("yolov8n.pt") # pass any model type model.train(epochs=5) ``` === "From scratch" ```python from ultralytics import YOLO model = YOLO("yolov8n.yaml") model.train(data="coco128.yaml", epochs=5) ``` === "Resume" ```python TODO: Resume feature is under development and should be released soon. ``` !!! example "Val" === "Val after training" ```python from ultralytics import YOLO model = YOLO("yolov8n.yaml") model.train(data="coco128.yaml", epochs=5) model.val() # It'll automatically evaluate the data you trained. ``` === "Val independently" ```python from ultralytics import YOLO model = YOLO("model.pt") # It'll use the data yaml file in model.pt if you don't set data. model.val() # or you can set the data you want to val model.val(data="coco128.yaml") ``` !!! example "Predict" === "From source" ```python from ultralytics import YOLO from PIL import Image import cv2 model = YOLO("model.pt") # accepts all formats - image/dir/Path/URL/video/PIL/ndarray. 0 for webcam results = model.predict(source="0") results = model.predict(source="folder", show=True) # Display preds. Accepts all YOLO predict arguments # from PIL im1 = Image.open("bus.jpg") results = model.predict(source=im1, save=True) # save plotted images # from ndarray im2 = cv2.imread("bus.jpg") results = model.predict(source=im2, save=True, save_txt=True) # save predictions as labels # from list of PIL/ndarray results = model.predict(source=[im1, im2]) ``` === "Results usage" ```python # results would be a list of Results object including all the predictions by default # but be careful as it could occupy a lot memory when there're many images, # especially the task is segmentation. # 1. return as a list results = model.predict(source="folder") # results would be a generator which is more friendly to memory by setting stream=True # 2. return as a generator results = model.predict(source=0, stream=True) for result in results: # detection result.boxes.xyxy # box with xyxy format, (N, 4) result.boxes.xywh # box with xywh format, (N, 4) result.boxes.xyxyn # box with xyxy format but normalized, (N, 4) result.boxes.xywhn # box with xywh format but normalized, (N, 4) result.boxes.conf # confidence score, (N, 1) result.boxes.cls # cls, (N, 1) # segmentation result.masks.masks # masks, (N, H, W) result.masks.segments # bounding coordinates of masks, List[segment] * N # classification result.probs # cls prob, (num_class, ) # Each result is composed of torch.Tensor by default, # in which you can easily use following functionality: result = result.cuda() result = result.cpu() result = result.to("cpu") result = result.numpy() ``` !!! note "Export and Deployment" === "Export, Fuse & info" ```python from ultralytics import YOLO model = YOLO("model.pt") model.fuse() model.info(verbose=True) # Print model information model.export(format=) # TODO: ``` === "Deployment" More functionality coming soon To know more about using `YOLO` models, refer Model class Reference [Model reference](reference/model.md){ .md-button .md-button--primary} --- ### Using Trainers `YOLO` model class is a high-level wrapper on the Trainer classes. Each YOLO task has its own trainer that inherits from `BaseTrainer`. !!! tip "Detection Trainer Example" ```python from ultralytics.yolo import v8 import DetectionTrainer, DetectionValidator, DetectionPredictor # trainer trainer = DetectionTrainer(overrides={}) trainer.train() trained_model = trainer.best # Validator val = DetectionValidator(args=...) val(model=trained_model) # predictor pred = DetectionPredictor(overrides={}) pred(source=SOURCE, model=trained_model) # resume from last weight overrides["resume"] = trainer.last trainer = detect.DetectionTrainer(overrides=overrides) ``` You can easily customize Trainers to support custom tasks or explore R&D ideas. Learn more about Customizing `Trainers`, `Validators` and `Predictors` to suit your project needs in the Customization Section. [Customization tutorials](engine.md){ .md-button .md-button--primary}