The limits of bio-molecular modeling with large language models : a cross-scale evaluation

arXiv cs.LG / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLMs’ effectiveness in bio-molecular discovery is not well established across multi-scale biological problems, motivating a more rigorous evaluation approach.
  • It introduces BioMol-LLM-Bench, a unified cross-scale benchmark with 26 downstream tasks spanning four difficulty levels, with computational tools integrated to assess tool-augmented capabilities.
  • Across evaluations of 13 representative models, the study finds that chain-of-thought prompting gives limited or negative benefit on biological tasks.
  • The results show hybrid mamba-attention architectures outperform for long bio-molecular sequences, while supervised fine-tuning increases specialization but can reduce generalization.
  • The authors conclude that current LLMs tend to do relatively well on classification but struggle on difficult regression tasks, offering guidance for future bio-molecular LLM modeling.

Abstract

The modeling of bio-molecular system across molecular scales remains a central challenge in scientific research. Large language models (LLMs) are increasingly applied to bio-molecular discovery, yet systematic evaluation across multi-scale biological problems and rigorous assessment of their tool-augmented capabilities remain limited. We reveal a systematic gap between LLM performance and mechanistic understanding through the proposed cross-scale bio-molecular benchmark: BioMol-LLM-Bench, a unified framework comprising 26 downstream tasks that covers 4 distinct difficulty levels, and computational tools are integrated for a more comprehensive evaluation. Evaluation on 13 representative models reveals 4 main findings: chain-of-thought data provides limited benefit and may even reduce performance on biological tasks; hybrid mamba-attention architectures are more effective for long bio-molecular sequences; supervised fine-tuning improves specialization at the cost of generalization; and current LLMs perform well on classification tasks but remain weak on challenging regression tasks. Together, these findings provide practical guidance for future LLM-based modeling of molecular systems.