Pragma-VL: Towards a Pragmatic Arbitration of Safety and Helpfulness in MLLMs
arXiv cs.LG / 3/17/2026
📰 NewsModels & Research
Key Points
- Pragma-VL proposes an end-to-end alignment scheme for multimodal LLMs to pragmatically arbitrate between safety and usefulness, addressing the safety-utility trade-off.
- It adds a cold-start supervised fine-tuning stage that improves visual risk perception via risk-aware clustering of the visual encoder and an interleaved dataset of risk descriptions and high-quality data.
- The approach introduces a theoretically-guaranteed reward model trained with a novel data augmentation method that assigns dynamic weights based on user queries to enable contextual arbitration between safety and helpfulness.
- Experimental results show Pragma-VL outperforms baselines by 5% to 20% on most multimodal safety benchmarks while preserving core capabilities in mathematics and knowledge reasoning.
Related Articles

Math needs thinking time, everyday knowledge needs memory, and a new Transformer architecture aims to deliver both
THE DECODER
Kreuzberg v4.5.0: We loved Docling's model so much that we gave it a faster engine
Reddit r/LocalLLaMA
Today, what hardware to get for running large-ish local models like qwen 120b ?
Reddit r/LocalLLaMA
Running mistral locally for meeting notes and it's honestly good enough for my use case
Reddit r/LocalLLaMA
[D] Single-artist longitudinal fine art dataset spanning 5 decades now on Hugging Face — potential applications in style evolution, figure representation, and ethical training data
Reddit r/MachineLearning