LLM-Based Data Generation and Clinical Skills Evaluation for Low-Resource French OSCEs

arXiv cs.CL / 4/10/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The paper proposes an LLM-driven pipeline to generate synthetic French OSCE doctor–patient dialogue transcripts in a low-resource setting where real annotated data is scarce.
  • It also uses LLM-assisted “silver labeling” to automatically evaluate the dialogues against scenario-specific clinical skills criteria, including a mix of ideal and perturbed performances to emulate different student proficiency levels.
  • Benchmarking across several open-source and proprietary LLMs finds that mid-size models (≤32B parameters) can reach accuracy on synthetic data comparable to GPT-4o (~90%), suggesting strong practical feasibility.
  • The authors argue the approach could enable privacy-preserving, locally deployable training-time evaluation systems that reduce reliance on human examiners for repeated practice and feedback in French medical education.
  • The work is positioned as a controlled way to create reproducible benchmarks for French OSCE assessment research despite the lack of real annotated transcripts.

Abstract

Objective Structured Clinical Examinations (OSCEs) are the standard method for assessing medical students' clinical and communication skills through structured patient interviews. In France, however, the organization of training sessions is limited by human and logistical constraints, restricting students' access to repeated practice and structured feedback. Recent advances in Natural Language Processing (NLP) and Large Language Models (LLMs) now offer the opportunity to automatically evaluate such medical interviews, thereby alleviating the need for human examiners during training. Yet, real French OSCE annotated transcripts remain extremely scarce, limiting reproducible research and reliable benchmarking. To address these challenges, we investigate the use of LLMs for both generating and evaluating French OSCE dialogues in a low-resource context. We introduce a controlled pipeline that produces synthetic doctor-patient interview transcripts guided by scenario-specific evaluation criteria, combining ideal and perturbed performances to simulate varying student skill levels. The resulting dialogues are automatically silver-labeled through an LLM-assisted framework supporting adjustable evaluation strictness. Benchmarking multiple open-source and proprietary LLMs shows that mid-size models (\le32B parameters) achieve accuracies comparable to GPT-4o (\sim90\%) on synthetic data, highlighting the feasibility of locally deployable, privacy-preserving evaluation systems for medical education.

LLM-Based Data Generation and Clinical Skills Evaluation for Low-Resource French OSCEs | AI Navigate