When Minor Edits Matter: LLM-Driven Prompt Attack for Medical VLM Robustness in Ultrasound
arXiv cs.CV / 3/24/2026
📰 News
Key Points
- The paper argues that vision-language models in ultrasound can be vulnerable to “prompt attacks” because even minor changes to natural-language instructions (typos, shorthand, ambiguity) can significantly alter outputs.
- It introduces a scalable adversarial evaluation framework that uses an LLM to generate clinically plausible, human-like prompt variants through minimal edits and “humanized” rewrites.
- The authors evaluate multiple state-of-the-art Med-VLMs on ultrasound multiple-choice question answering benchmarks to measure vulnerability, including how attacker model capability affects success rates.
- They analyze how attack success correlates with model confidence and report consistent failure patterns across models, indicating realistic robustness gaps for safe clinical deployment.
- The authors plan to release the code publicly after the review process, enabling further testing and mitigation work.
- categories: [