FLARE: Fully Integration of Vision-Language Representations for Deep Cross-Modal Understanding

arXiv cs.CV / 4/30/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • FLARE is a new family of vision-language models that performs full vision-language alignment and deep integration across the entire pipeline, rather than using simple MLP projectors and leaving cross-modal interaction to later LLM decoding.
  • The approach includes text-guided vision encoding for pixel-level alignment, context-aware alignment decoding that aggregates visual features conditioned on text, and a dual-semantic mapping loss to bridge representations between modalities.
  • FLARE also uses text-driven VQA synthesis to generate high-quality VQA pairs and corresponding images for data-level optimization.
  • Experiments train FLARE at 3B and 8B scales (with fixed and dynamic resolutions), where FLARE shows strong performance improvements over prior methods, including beating larger baselines like Cambrian-1 8B and Florence-VL 8B, and maintaining generalizability.
  • The authors provide released code, model weights, and a dataset via the project repository, enabling replication and further research.

Abstract

We introduce FLARE, a family of vision language models (VLMs) with a fully vision-language alignment and integration paradigm. Unlike existing approaches that rely on single MLP projectors for modality alignment and defer cross-modal interaction to LLM decoding, FLARE achieves deep, dynamic integration throughout the pipeline. Our key contributions include: (1) Text-Guided Vision Encoding that incorporates textual information during vision encoding to achieve pixel-level alignment; (2) Context-Aware Alignment Decoding that aggregates visual features conditioned on textual context during decoding for query-level integration; (3) Dual-Semantic Mapping Loss to supervise feature mapping from both modalities and enable modality-level bridging; and (4) Text-Driven VQA Synthesis that leverages high-quality text to generate VQA pairs and synthesize corresponding images, enabling data-level optimization. We train FLARE at 3B and 8B scales under both fixed and dynamic resolution settings, demonstrating that our full-modality alignment significantly outperforms existing methods while maintaining strong generalizability. FLARE 3B surpasses Cambrian-1 8B and Florence-VL 8B using only 630 vision tokens. Ablation studies reveal that FLARE achieves superior performance over existing methods with minimal computational cost. Even without dynamic resolution, FLARE outperforms LLaVA-NeXT, validating the effectiveness of our approach. We release our code, model weights, and dataset in https://github.com/starriver030515/FLARE.