Skip to content

YOLOv8 - Keypoint Detection

YOLOv8 is a popular computer vision model developed by Ultralytics. The keypoint detection variant (also known as pose estimation) detects objects and predicts keypoints for each detected instance, commonly used for human pose estimation.

Overview

YOLOv8 for keypoint detection combines object detection with keypoint localization. Key features include:

  • Anchor-free detection head - Simplified architecture without anchor boxes
  • Keypoint prediction - Predicts x, y coordinates and confidence for each keypoint
  • Multi-person support - Detects and tracks keypoints for multiple people simultaneously
  • Improved feature pyramid network - Better multi-scale feature fusion
  • Multiple model sizes - From nano to extra-large variants

License

AGPL-3.0

Commercial Licensing

  • AGPL-3.0: Free for open-source projects. Requires derivative works to be open-sourced.
  • Paid Roboflow customers: Automatically get access to use any YOLOv8 models trained on or uploaded to the Roboflow platform for commercial use.
  • Free Roboflow customers: Can use YOLOv8 via the serverless hosted API, or commercially self-hosted with a paid plan.

Learn more: Roboflow Licensing | YOLOv8 License Details

Pre-trained Model IDs

All pre-trained YOLOv8 keypoint detection models are trained on the COCO dataset (17 keypoints for human pose) and are open access (no API key required).

Model Size 640×640 1280×1280
Nano yolov8n-pose-640 -
Small yolov8s-pose-640 -
Medium yolov8m-pose-640 -
Large yolov8l-pose-640 -
Extra-Large yolov8x-pose-640 yolov8x-pose-1280

COCO Keypoints: nose, left_eye, right_eye, left_ear, right_ear, left_shoulder, right_shoulder, left_elbow, right_elbow, left_wrist, right_wrist, left_hip, right_hip, left_knee, right_knee, left_ankle, right_ankle

Supported Backends

Backend Extras Required
onnx onnx-cpu, onnx-cu12, onnx-cu118, onnx-jp6-cu126
torch-script torch-cpu, torch-cu118, torch-cu124, torch-cu126, torch-cu128, torch-jp6-cu126
trt trt10

Roboflow Platform Compatibility

Feature Supported
Training ✅ Train custom models on Roboflow
Upload Weights ✅ Upload pre-trained weights
Serverless API (v2) Deploy via hosted API
Workflows ✅ Use in Workflows via Keypoint Detection block
Edge Deployment (Jetson) ✅ Deploy on NVIDIA Jetson devices
Self-Hosting ✅ Deploy with inference-models

Custom model ID format: project-url/version (e.g., my-project-abc123/2)

Usage Example

import cv2
import supervision as sv
from inference_models import AutoModel

# Load model and image
model = AutoModel.from_pretrained("yolov8n-pose-640")
image = cv2.imread("path/to/image.jpg")

# Run inference - returns (List[KeyPoints], Optional[List[Detections]])
results = model(image)
key_points_list, detections_list = results

# Convert to supervision format for filtering and visualization
key_points = key_points_list[0].to_supervision()

# Filter for "person" class (class_id=0) for COCO pose models
person_mask = key_points.class_id == 0
key_points_filtered = key_points[person_mask]

# Annotate image with keypoints
vertex_annotator = sv.VertexAnnotator()
edge_annotator = sv.EdgeAnnotator(edges=model.skeletons[0])

annotated_image = edge_annotator.annotate(image.copy(), key_points_filtered)
annotated_image = vertex_annotator.annotate(annotated_image, key_points_filtered)

# Save or display
cv2.imwrite("annotated.jpg", annotated_image)

Output Format

Returns: Tuple[List[KeyPoints], Optional[List[Detections]]] (one element per image in batch)

Skeleton connections: Access via model.skeletons[class_id]. The model.skeletons list has one element per detection class. For COCO pose models (single "person" class), use model.skeletons[0]. For custom multi-class models, each class has its own skeleton at the corresponding index.

Pass to EdgeAnnotator: sv.EdgeAnnotator(edges=model.skeletons[0]) or omit for auto-detection.