Seq vs Seq: An Open Suite of Paired Encoders and Decoders
arXiv cs.CL / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The authors introduce the Ettin suite—paired encoder-only and decoder-only models from 17M to 1B parameters trained on up to 2 trillion tokens—and show SOTA recipes for both categories by using the same training recipe, beating ModernBERT for encoders and Llama 3.2 and SmolLM2 for decoders.
Related Articles

How to Build an AI Team: The Solopreneur Playbook
Dev.to

CrewAI vs AutoGen vs LangGraph: Which Agent Framework to Use
Dev.to

14 Best Self-Hosted Claude Alternatives for AI and Coding in 2026
Dev.to
[P] Finetuned small LMs to VLM adapters locally and wrote a short article about it
Reddit r/MachineLearning
Experiment: How far can a 28M model go in business email generation?
Reddit r/LocalLLaMA