RXNRECer Enables Fine-grained Enzymatic Function Annotation through Active Learning and Protein Language Models
arXiv cs.LG / 3/16/2026
📰 NewsModels & Research
Key Points
- The paper introduces RXNRECer, a transformer-based ensemble that directly predicts enzyme-catalyzed reactions without relying on EC numbers, addressing ambiguities from EC mappings.
- It combines protein language modeling with active learning to capture both high-level sequence semantics and fine-grained transformation patterns.
- On curated cross-validation and temporal test sets, RXNRECer shows gains of 16.54% in F1 score and 15.43% in accuracy over six EC-based baselines, enabling scalable proteome-wide reaction annotation.
- The framework provides interpretable prediction rationales via large language models and has broad potential applications in enzyme research and industrial contexts.
Related Articles

Math needs thinking time, everyday knowledge needs memory, and a new Transformer architecture aims to deliver both
THE DECODER
Kreuzberg v4.5.0: We loved Docling's model so much that we gave it a faster engine
Reddit r/LocalLLaMA
Today, what hardware to get for running large-ish local models like qwen 120b ?
Reddit r/LocalLLaMA
Running mistral locally for meeting notes and it's honestly good enough for my use case
Reddit r/LocalLLaMA
[D] Single-artist longitudinal fine art dataset spanning 5 decades now on Hugging Face — potential applications in style evolution, figure representation, and ethical training data
Reddit r/MachineLearning