FusionAgent: A Multimodal Agent with Dynamic Model Selection for Human Recognition

arXiv cs.CV / 3/31/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • FusionAgent is an agentic multimodal framework for whole-body human recognition that performs dynamic, sample-specific model selection instead of static score fusion across all inputs.
  • It treats each expert model as a tool and uses Reinforcement Fine-Tuning with a metric-based reward to learn the optimal model combination per test sample.
  • To improve fusion quality under score misalignment and embedding heterogeneity, it introduces Anchor-based Confidence Top-k (ACT) score fusion that anchors on the most confident model and fuses complementary predictions in a confidence-aware way.
  • Experiments on multiple whole-body biometric benchmarks report state-of-the-art performance with higher efficiency, attributed to fewer model invocations.
  • The work emphasizes dynamic, explainable, and robust model fusion as a key ingredient for real-world recognition in unconstrained settings.

Abstract

Model fusion is a key strategy for robust recognition in unconstrained scenarios, as different models provide complementary strengths. This is especially important for whole-body human recognition, where biometric cues such as face, gait, and body shape vary across samples and are typically integrated via score-fusion. However, existing score-fusion strategies are usually static, invoking all models for every test sample regardless of sample quality or modality reliability. To overcome these limitations, we propose \textbf{FusionAgent}, a novel agentic framework that leverages a Multimodal Large Language Model (MLLM) to perform dynamic, sample-specific model selection. Each expert model is treated as a tool, and through Reinforcement Fine-Tuning (RFT) with a metric-based reward, the agent learns to adaptively determine the optimal model combination for each test input. To address the model score misalignment and embedding heterogeneity, we introduce Anchor-based Confidence Top-k (ACT) score-fusion, which anchors on the most confident model and integrates complementary predictions in a confidence-aware manner. Extensive experiments on multiple whole-body biometric benchmarks demonstrate that FusionAgent significantly outperforms SoTA methods while achieving higher efficiency through fewer model invocations, underscoring the critical role of dynamic, explainable, and robust model fusion in real-world recognition systems. Project page: \href{https://fusionagent.github.io/}{FusionAgent}.