Automatic Image-Level Morphological Trait Annotation for Organismal Images

arXiv cs.CV / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a scalable pipeline for automatically annotating organism images with morphological traits, addressing the lack of high-quality image-to-trait datasets that slows expert-driven annotation.
  • It shows that sparse autoencoders trained on foundation-model features can produce monosemantic, spatially grounded neurons that reliably respond to meaningful morphological parts.
  • The method localizes salient regions and uses vision-language prompting to generate interpretable trait descriptions, then validates them for biological plausibility via human evaluation.
  • It introduces Bioscan-Traits, a new dataset containing 80K trait annotations across 19K insect images derived from BIOSCAN-5M.
  • The authors perform an extensive ablation study to measure how sensitive the trait-description quality is to key design choices, aiming to provide guidance for effective deployment of the pipeline.

Abstract

Morphological traits are physical characteristics of biological organisms that provide vital clues on how organisms interact with their environment. Yet extracting these traits remains a slow, expert-driven process, limiting their use in large-scale ecological studies. A major bottleneck is the absence of high-quality datasets linking biological images to trait-level annotations. In this work, we demonstrate that sparse autoencoders trained on foundation-model features yield monosemantic, spatially grounded neurons that consistently activate on meaningful morphological parts. Leveraging this property, we introduce a trait annotation pipeline that localizes salient regions and uses vision-language prompting to generate interpretable trait descriptions. Using this approach, we construct Bioscan-Traits, a dataset of 80K trait annotations spanning 19K insect images from BIOSCAN-5M. Human evaluation confirms the biological plausibility of the generated morphological descriptions. We assess design sensitivity through a comprehensive ablation study, systematically varying key design choices and measuring their impact on the quality of the resulting trait descriptions. By annotating traits with a modular pipeline rather than prohibitively expensive manual efforts, we offer a scalable way to inject biologically meaningful supervision into foundation models, enable large-scale morphological analyses, and bridge the gap between ecological relevance and machine-learning practicality.