Keypoint Annotation: The Complete Technical Guide to Tools, COCO Format, and Services (2026)

The definitive guide to keypoint annotation for ML teams. Covers COCO keypoint format, CVAT keypoint annotation workflow, free tools comparison, use cases in pose estimation and robotics — plus when to outsource to a professional keypoint annotation service.

Bounding Boxes Were Never Enough.

A bounding box tells your model that a person exists in a frame. That’s it.

It cannot tell your model how that person is standing. It cannot tell a surgical robot where to grip. It cannot tell a sports analytics system whether the athlete’s knee angle increases injury risk. It cannot tell a face authentication system which facial micro-expressions belong to a live human versus a spoofing attack.

That’s the problem keypoint annotation solves — and it’s why demand for precise keypoint annotation services has accelerated sharply across robotics, healthcare, biometrics, augmented reality, and human pose estimation.

If you’re an ML engineer, CV researcher, or AI Founder building any model that needs to understand structure, pose, or movement — this guide covers everything: what keypoint annotation is, the COCO format, the best tools (free and paid), real-world use cases, and what to look for when outsourcing to a keypoint annotation service.


What Is Keypoint Annotation?

Keypoint annotation is the process of marking specific, precisely defined coordinate points on objects in images or video frames. Each marked point — called a landmark or keypoint — represents a meaningful structural location: a joint on a human body, a corner of an eye, a mechanical pivot on a robotic part, or a reference point on a vehicle.

Unlike bounding boxes (which capture presence) or segmentation masks (which capture shape), keypoints capture geometry and spatial relationships — the internal architecture of the object.

A trained model learns not just where an object is, but how it’s configured. This distinction is what separates models that can detect something from models that can understand it.

Common keypoint annotation schemas:

  • Human body: 17-point COCO skeleton (wrists, elbows, shoulders, hips, knees, ankles, eyes, ears, nose)
  • Human face/facial landmarks: 68-point or 98-point schema for facial recognition and liveness detection
  • Hand: 21-point hand skeleton for gesture recognition and sign language AI
  • Vehicle: Corner points, wheel centers, sensor mount positions
  • Animal: Species-specific skeletal schemas for wildlife monitoring or veterinary AI
  • Industrial parts: Custom grasping points, bolt holes, alignment markers for robotic manipulation

The COCO Keypoint Annotation Format

The COCO (Common Objects in Context) keypoint format is the industry standard for human pose estimation and the most widely supported keypoint schema across ML frameworks. If you’re building a pose estimation model with YOLOv8-pose, MMPose, or any other modern architecture — you’ll encounter COCO keypoint format.

The COCO keypoint schema (17 points):

0: nose
1: left_eye
2: right_eye
3: left_ear
4: right_ear
5: left_shoulder
6: right_shoulder
7: left_elbow
8: right_elbow
9: left_wrist
10: right_wrist
11: left_hip
12: right_hip
13: left_knee
14: right_knee
15: left_ankle
16: right_ankle

How keypoints are stored in COCO JSON format:

Each annotation object includes a keypoints field — a flat list of [x, y, visibility] triplets for each point:

{
  "id": 1,
  "image_id": 101,
  "category_id": 1,
  "keypoints": [
    320, 240, 2,
    310, 225, 2,
    330, 225, 2,
    295, 220, 1,
    345, 220, 0,
    ...
  ],
  "num_keypoints": 14,
  "bbox": [280, 180, 140, 320],
  "area": 44800,
  "iscrowd": 0
}

Visibility flags — the detail most annotators get wrong:

Value Meaning
0 Not labeled (point not annotated)
1 Labeled but not visible (occluded or out-of-frame — coordinates are estimated)
2 Labeled and visible (annotator can clearly see the point)

This visibility flag is not optional noise. Models trained with correct visibility labeling learn to handle occlusion gracefully. Models trained without it — or with incorrect flags — produce jittery, unreliable keypoint predictions in production, especially in crowded scenes.

YOLOv8-pose keypoint format:

For YOLOv8 pose training specifically, keypoints are stored in YOLO TXT format with normalized coordinates:

<class_id> <x_center> <y_center> <w> <h> <kp1_x> <kp1_y> <kp1_vis> <kp2_x> <kp2_y> <kp2_vis> ...

All coordinates normalized between 0 and 1. Each person gets one line. All 17 COCO keypoints listed in order.


Keypoint Annotation Tools — Free and Paid

CVAT (Free, Open-Source) — Best for Teams

CVAT keypoint annotation is the gold standard for professional teams running structured annotation workflows. CVAT supports the full COCO 17-point skeleton schema natively and lets you define custom keypoint schemas for non-human use cases.

How to annotate keypoints in CVAT:

  1. Create a project → add a label → set label type to Skeleton.
  2. Define your keypoint schema by adding individual point nodes and connecting edges (e.g., connecting “left_shoulder” to “left_elbow”).
  3. In the annotation editor, select the skeleton tool and click to place each keypoint.
  4. Set visibility states (visible / occluded / outside frame) per point.
  5. Export in COCO Keypoints format or YOLO Pose format.

CVAT’s interpolation feature is critical for video: annotate the keyframe skeleton, set interpolation mode, and CVAT propagates the skeleton across frames. Human reviewers then correct tracking errors. This workflow is 5–8x faster than frame-by-frame manual annotation on video datasets.

CVAT keypoint annotation: best for — team workflows, video datasets, self-hosted data privacy requirements, and COCO-format output.


Roboflow (Free tier available) — Best for Speed

Roboflow supports keypoint annotation with SAM-2-powered assistance and direct export to YOLO Pose format. The auto-labeling tools can accelerate initial keypoint placement on common human pose tasks, though human QA remains mandatory for precision applications.

Best for — solo ML engineers, startup MVPs, and teams that need fast annotation with direct YOLO Pose export.


Label Studio (Free, Open-Source) — Best for Custom Schemas

Label Studio supports fully custom keypoint schemas via its XML template system. If you’re annotating something other than human pose — industrial parts, animal skeletons, custom facial landmark schemas — Label Studio gives you the most flexibility in defining your own keypoint taxonomy.

Best for — research teams, non-standard object types, multi-modal projects where keypoint annotation is one of several data types.


Supervisely (Free tier, Enterprise) — Best for Scale

Supervisely has one of the strongest keypoint annotation interfaces in the market, with built-in smart tooling and team management. Strong choice when you’re running keypoint annotation at scale with multiple annotators and need robust QA workflows.

Best for — larger teams, mixed annotation types alongside keypoints, enterprise data governance requirements.


Keypoint Annotation Use Cases in 2026

Human Pose Estimation

The most established use case. Keypoint annotation trains models to detect the human body skeleton across a range of applications:

  • Fitness and rehabilitation AI — Detecting joint angles in real time to identify incorrect exercise form or track physical therapy recovery.
  • Sports analytics — Measuring stride length, shoulder rotation, and knee flexion for professional athletes. The same models flag injury risk patterns.
  • Safety compliance — Detecting worker posture in industrial environments to identify ergonomic risks before they become injuries.
  • Action recognition — Classifying activity type (running, jumping, falling) from skeleton trajectory, not just pixel appearance.

Facial Landmark Detection

The most precision-demanding keypoint annotation task. A 68-point facial landmark schema — covering the jaw line, eyebrows, eyes, nose, and mouth — powers:

  • Face recognition and verification — Alignment preprocessing before running identity matching. Misaligned keypoints degrade matching accuracy directly.
  • Biometric liveness detection — Tracking micro-expressions and facial geometry changes to distinguish live humans from 2D/3D spoofing attacks. This is the 2026 anti-spoofing standard for high-security banking and government applications.
  • Emotion recognition — Classifying emotional states from the geometry of facial muscle movements.
  • AR/VR face filters — Aligning digital overlays to precise facial geometry in real time.

Robotics and Industrial Automation

This is the emerging frontier for keypoint annotation beyond human pose:

  • Robotic grasping — Labeling specific grip points on irregular industrial components so robotic arms can grasp with sub-millimeter accuracy.
  • Assembly verification — Keypointing bolt holes, alignment marks, and component edges to train visual inspection models.
  • Human-robot collaboration (cobots) — Tracking human hand and body keypoints in shared workspaces to prevent collisions.

Medical Imaging

  • Anatomical landmark annotation — Marking bone landmarks in X-rays and CT scans for surgical planning and orthopedic AI.
  • Tumor boundary reference points — Keypointing reference anatomy around lesion sites for measurement consistency across imaging sessions.

Sign Language and Gesture AI

  • Hand keypoint annotation — 21-point hand skeleton labeling for sign language recognition models. Each of the 21 points maps to a specific joint in the hand and fingers.
  • Gesture control interfaces — Training models that interpret specific hand configurations as commands.

What Makes Keypoint Annotation Hard (And Expensive to Get Wrong)

Keypoint annotation has higher precision requirements than any other annotation type. The margin for error is measured in pixels — and pixel-level errors compound across training.

The five failure modes that wreck keypoint-trained models:

  1. Incorrect visibility flagging. An annotator marks an occluded knee as “visible” because they can infer its position. The model learns that it should produce confident predictions for invisible joints — the exact opposite of what you want.

  2. Inconsistent anatomical convention. Two annotators have different intuitions about where exactly the “shoulder keypoint” sits — the top of the deltoid, the glenohumeral joint center, or the acromion. Without a strict labeling guide with reference images, your dataset will have two different “shoulders” and your model will learn a noisy average of both.

  3. Temporal jitter in video. A keypoint that jumps 5–10 pixels between consecutive frames where the subject hasn’t moved. Models trained on jittery video annotations produce jittery predictions — a critical failure in surgical and rehabilitation applications where temporal smoothness is a hard requirement.

  4. Missing edge case coverage. No annotations at extreme joint angles (arms fully extended overhead, deep squat positions, lateral bends). The model will fail exactly at the positions that matter most for athletic performance or injury detection.

  5. Schema drift. Annotators progressively shift their interpretation of a keypoint placement convention over a long project without anyone catching it. The first 500 images use one convention; the last 500 use a slightly different one. Your training data contains a silent, invisible inconsistency.

These failures are invisible in your annotation tool. They only surface when model performance degrades in production — often weeks after the dataset was “complete.”


In-House vs. Outsourced Keypoint Annotation: The Real Decision

Be honest with yourself about what in-house keypoint annotation actually costs.

Your ML engineers spend their time debugging annotation inconsistencies, writing labeling guides, and reviewing annotator output instead of building model architecture and training pipelines. That’s your highest-cost resource doing your lowest-leverage work.

Outsource when:

  • Your dataset requires 500+ annotated keypoint instances.
  • You need temporal consistency across video frames.
  • Your annotation schema is complex (21+ points per instance).
  • Your use case is precision-critical (medical, biometrics, surgical robotics).
  • You have a deadline and a team that’s already stretched.

Keep in-house when:

  • You’re prototyping with fewer than 100 images.
  • You have a simple 5–7 point custom schema with very high self-familiarity.
  • You need total data confidentiality with zero third-party access.

Why Keypoint Annotation Quality Directly Controls Model Ceiling

Improving annotation quality from 85% to 97% accuracy on a keypoint dataset has a measurable impact on model performance that exceeds most architecture upgrades at equivalent cost. This is not a theoretical claim — it’s the operational reality of data-centric AI.

The coordinate precision of your keypoints sets the upper bound of your model’s localization accuracy. You cannot train above your label quality.

At AI and ML Network, keypoint annotation is one of our core specializations. We work daily with COCO 17-point human pose schemas, custom facial landmark schemas (68-point and 98-point), hand skeleton annotation, and industrial part keypointing. Every project includes:

  • A strict labeling guide with reference images before annotation begins.
  • Visibility flag review as part of the QA pass — not just coordinate review.
  • Temporal consistency checking on all video datasets.
  • COCO JSON and YOLO Pose format export, ready for training.

We work at affordable rates compared to other keypoint annotation services, and we maintain accuracy standards that give your model a clean foundation to train from.

Need a free 50-image sample batch? Send us your keypoint schema and we’ll annotate a sample batch — COCO format, visibility flags included. Judge the quality before committing to anything.

Contact AI and ML Network


Alt text for cover image: Diagram showing COCO 17-point keypoint annotation schema on a human body with visibility flags, alongside CVAT annotation interface for keypoint labeling in 2026