AI Navigate

Pragma-VL: Towards a Pragmatic Arbitration of Safety and Helpfulness in MLLMs

arXiv cs.LG / 3/17/2026

📰 NewsModels & Research

Key Points

  • Pragma-VL proposes an end-to-end alignment scheme for multimodal LLMs to pragmatically arbitrate between safety and usefulness, addressing the safety-utility trade-off.
  • It adds a cold-start supervised fine-tuning stage that improves visual risk perception via risk-aware clustering of the visual encoder and an interleaved dataset of risk descriptions and high-quality data.
  • The approach introduces a theoretically-guaranteed reward model trained with a novel data augmentation method that assigns dynamic weights based on user queries to enable contextual arbitration between safety and helpfulness.
  • Experimental results show Pragma-VL outperforms baselines by 5% to 20% on most multimodal safety benchmarks while preserving core capabilities in mathematics and knowledge reasoning.

Abstract

Multimodal Large Language Models (MLLMs) pose critical safety challenges, as they are susceptible not only to adversarial attacks such as jailbreaking but also to inadvertently generating harmful content for benign users. While internal safety alignment via Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) is a primary mitigation strategy, current methods often face a safety-utility trade-off: they either refuse benign queries out of excessive caution or overlook latent risks in cross-modal interactions. To resolve this, we introduce Pragma-VL, an end-to-end alignment algorithm that enables MLLMs to pragmatically arbitrate between safety and helpfulness. First, we enhance visual risk perception with a novel cold-start SFT stage. This is achieved by applying risk-aware clustering to the visual encoder and using an interleaved dataset of risk descriptions and high-quality data. Second, we introduce a theoretically-guaranteed reward model that leverages synergistic learning. We train it with a novel data augmentation method that assigns dynamic weights based on the queries, enabling contextual arbitration between safety and helpfulness. Extensive experiments show that Pragma-VL effectively balances safety and helpfulness, outperforming baselines by 5% to 20% on most multimodal safety benchmarks while preserving its general capabilities in areas such as mathematics and knowledge reasoning.