internlm/Intern-S2-Preview · Hugging Face

Reddit r/LocalLLaMA / 5/15/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • Hugging Face introduces Intern-S2-Preview, an efficient 35B scientific multimodal foundation model that focuses on “task scaling” rather than only scaling parameters and data.
  • The model extends professional scientific tasks into a full-chain training pipeline (pre-training through reinforcement learning), achieving performance comparable to a trillion-scale Intern-S1-Pro on key professional scientific tasks while using only 35B parameters.
  • Intern-S2-Preview improves multimodal scientific capabilities, including stronger spatial modeling for small-molecule structures and real-valued prediction modules, and claims the first open-source offering with both material crystal-structure generation and strong general capabilities.
  • It enhances agent capabilities for scientific workflows and reports strong results across multiple scientific agent benchmarks.
  • During reinforcement learning, the model uses shared-weight MTP with KL loss to reduce training–inference mismatch and improve MTP acceptance rate and token generation speed, along with CoT compression to shorten responses without sacrificing reasoning performance.
internlm/Intern-S2-Preview · Hugging Face

Introduction

We introduce Intern-S2-Preview, an efficient 35B scientific multimodal foundation model. Beyond conventional parameter and data scaling, Intern-S2-Preview explores task scaling: increasing the difficulty, diversity, and coverage of scientific tasks to further unlock model capabilities.

By extending professional scientific tasks into a full-chain training pipeline from pre-training to reinforcement learning, Intern-S2-Preview achieves performance comparable to the trillion-scale Intern-S1-Pro on multiple core professional scientific tasks, while using only 35B parameters (continued pretrained from Qwen3.5). At the same time, it maintains strong general reasoning, multimodal understanding, and agent capabilities.

Features

  • Scientific task scaling with full-chain training. Intern-S2-Preview scales hundreds of professional scientific tasks from pre-training to RL, enabling strong performance across multiple specialized domains at only 35B parameters. It further strengthens spatial modeling for small-molecule structures and introduces real-valued prediction modules, making it the first open-source model with both material crystal structure generation capability and strong general capabilities.
  • Enhanced agent capabilities for scientific workflows. Intern-S2-Preview significantly improves agentic abilities over the previous generation, achieving strong results on multiple scientific agent benchmarks.
  • Efficient RL reasoning with MTP and CoT compression. During RL, Intern-S2-Preview adopts shared-weight MTP with KL loss to reduce the mismatch between training and inference behavior, substantially improving MTP accept rate and token generation speed. It also introduces CoT compression techniques to shorten responses while preserving strong reasoning capability, achieving improvements in both performance and efficiency.
submitted by /u/pmttyji
[link] [comments]