healthneutral

Standardising Surgical Movements for Smarter AI

Saturday, May 16, 2026

Recent research demonstrates that machines grasp surgical nuances far better when they focus on tiny, intentional actions—the gestures that occur as a tool touches tissue. These micro‑movements provide a clearer picture than broad labels like “cut” or “close.” By linking gestures to surgeon skill, AI can even predict patient outcomes.

The Problem: Fragmented Terminology

  • No common language: Hospitals and research groups use unique terms.
  • Data silos: Sharing datasets or comparing AI models becomes difficult.
  • Stalled progress: The lack of standardization hampers the development of reliable, generalizable surgical tools.

The Solution: A Consensus‑Driven Taxonomy

A consortium from the Society of American Gastrointestinal and Endoscopic Surgeons (SAGES) employed a Delphi method—a structured round‑robin survey—to gather surgeon and researcher input. The result: a unified set of gesture definitions that:

  • Standardizes datasets across institutions.
  • Facilitates reproducibility of AI models.
  • Enables cross‑hospital, cross‑specialty training.

Impact on Surgical AI

  • Broader applicability: Models trained with the taxonomy work in varied operating rooms.
  • Performance measurement: Future studies can assess how mastering new gestures improves surgeon skill.
  • Path to real tools: A shared framework is essential for translating research into safer, more effective surgical technology.

By agreeing on what constitutes a gesture, the surgical AI community moves from scattered vocabularies to a common framework—paving the way for tangible improvements in operating room safety and patient outcomes.

Actions