LoFi: Location-Aware Fine-Grained Representation Learning for Chest X-ray

arXiv cs.AI / 3/23/2026

💬 OpinionModels & Research

Key Points

  • LoFi introduces a location-aware fine-grained representation learning framework for chest X-rays that uses region-level supervision via a location-aware captioning loss to improve grounding and dense captioning.
  • The approach jointly optimizes sigmoid, captioning, and location-aware captioning losses using a lightweight large language model to learn fine-grained, region-specific representations.
  • A fine-grained encoder is integrated into retrieval-based in-context learning to enhance chest X-ray grounding across diverse clinical settings.
  • Experiments on MIMIC-CXR and PadChest-GR demonstrate superior retrieval and phrase grounding performance, highlighting practical improvements in fine-grained medical image understanding.

Abstract

Fine-grained representation learning is crucial for retrieval and phrase grounding in chest X-rays, where clinically relevant findings are often spatially confined. However, the lack of region-level supervision in contrastive models and the limited ability of large vision language models to capture fine-grained representations in external validation lead to suboptimal performance on these tasks. To address these limitations, we propose Location-aware Fine-grained representation learning (LoFi), which jointly optimizes sigmoid, captioning, and location-aware captioning losses using a lightweight large language model. The location-aware captioning loss enables region-level supervision through grounding and dense captioning objectives, thereby facilitating fine-grained representation learning. Building upon these representations, we integrate a fine-grained encoder into retrieval-based in-context learning to enhance chest X-ray grounding across diverse settings. Extensive experiments demonstrate that our method achieves superior retrieval and phrase grounding performance on MIMIC-CXR and PadChest-GR.