ClinicalSim.ai
Back to Insights
·6 min read·ClinicalSim.ai Team

What Medical Learners Actually Want from AI Standardized Patients

New CHI 2026 research reveals six key requirements for AI-SP design—straight from the medical students who would use them.

The Study

A research team from CUHK Shenzhen, Nankai University, and MIT Media Lab spent months with clinical-year medical students. They wanted to know: what would actually make AI standardized patients useful?

They interviewed 12 students. Ran three co-design workshops. The findings don't match what most vendors assume about medical simulation.

The Problem

The paper's title says it all: "It Talks Like a Patient, But Feels Different."

Current AI systems generate natural-sounding dialogue. But students describe the experience as flat. Something essential is missing.

This isn't a technical problem. It's a design problem.

What Students Actually Asked For

1. Different Modes for Different Goals

Students don't want one generic AI patient. They want options:

  • OSCE mode: Rigid, standardized, exam-like encounters
  • Practice mode: Open-ended scenarios where they can explore
  • Skill-building mode: Scaffolded practice with adjustable difficulty

The simulation should match what you're trying to learn. Not aim for some abstract "realism."

2. Clear Information Rules

Current AI patients feel like guessing games. Students want transparent rules:

  • Visual cues (jaundice, rashes) should be immediately apparent
  • Information should require appropriate clinical questions
  • Rules should stay consistent across encounters

3. Ways to Actually Examine Patients

Human actors can't safely do physical exams. Students wanted AI patients to fill this gap:

  • Body maps for visual examination
  • Sensory cues (sounds, images)
  • Virtual test ordering with realistic timing
  • Safe practice of procedures that would be dangerous on humans

The point isn't perfect sensory fidelity. It's practicing how communication and evidence-gathering work together.

4. Backup Options When Voice Fails

Voice-based AI breaks down. Students anticipated this:

  • Voice as the primary input method
  • Text fallback when speech recognition fails
  • Keyword shortcuts for specific queries
  • Hints when they get stuck

This prevents technical failures from derailing the learning experience.

5. Control Over Difficulty and Feedback

Students reframed AI patients as practice tools, not tests:

  • Adjustable difficulty from "helpful mode" to "stress-test mode"
  • Light feedback during encounters (trust indicators, facial expressions)
  • Structured review after sessions
  • Links from mistakes to textbook references

6. Variable Patient Personalities

Real patients differ wildly. Students wanted to practice with:

  • Selectable personas (elderly, pediatric, anxious)
  • Variable emotional responses (crying, frustration, relief)
  • Different cultural backgrounds and health beliefs
  • Replayable scenarios with the same case but different patient types

This makes affect a training variable you can control.

The Real Insight

Students don't judge AI patients by how natural the conversation sounds. They judge by whether the system helps them learn.

Natural conversation doesn't equal educational value.

What actually matters:

  • Can I tell what mode I'm in?
  • Can I adjust the difficulty?
  • Do I understand how I'm doing?
  • Does the feedback help me improve?

Where AI Patients Fit

The researchers position AI patients as complements to human actors:

Human SPs handle: high-stakes assessment, embodied presence, nuanced social interaction, expert feedback

AI patients handle: unlimited repetition, configurable difficulty, structured feedback, on-demand availability

Each tool does what it's best at.

What This Means for ClinicalSim.ai

This research validates some choices we made:

  • Voice-first with text backup matches what students wanted
  • Structured scenarios with clear objectives align with their needs
  • Immediate feedback addresses the gap they identified
  • On-demand availability fills the gap between limited human SP sessions

It also points to what's next: explicit difficulty modes, more patient personas, stronger post-session debriefs.

The Bottom Line

Medical students aren't looking for AI patients that perfectly replicate humans. They want reliable, transparent, controllable tools that let them practice high-stakes conversations without the logistical headaches.

The "empathy gap" isn't a reason to give up on AI patients. It's a design challenge: build for instructional usability, not conversational fluency.


Based on: Gao, Z., et al. (2026). "It Talks Like a Patient, But Feels Different": Co-Designing AI Standardized Patients with Medical Learners. CHI 2026, Barcelona, Spain.