MOSS-TTS Technical Report
arXiv cs.CL / 3/20/2026
📰 NewsModels & Research
Key Points
- MOSS-TTS is a speech generation foundation model built on a scalable recipe using discrete audio tokens, autoregressive modeling, and large-scale pretraining.
- It is built on MOSS-Audio-Tokenizer, a causal Transformer tokenizer that compresses 24 kHz audio to 12.5 fps with variable-bitrate RVQ and unified semantic-acoustic representations.
- The release includes two generators: MOSS-TTS, emphasizing structural simplicity, scalability, and long-context/control-oriented deployment, and MOSS-TTS-Local-Transformer, which adds a frame-local autoregressive module for higher efficiency, stronger speaker preservation, and a shorter time to first audio.
- Across multilingual and open-domain settings, MOSS-TTS supports zero-shot voice cloning, token-level duration control, phoneme-/pinyin-level pronunciation control, smooth code-switching, and stable long-form generation.
- The report summarizes the design, training recipe, and empirical characteristics of the released models.
Related Articles

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to
How to settle on a coding LLM ? What parameters to watch out for ?
Reddit r/LocalLLaMA

Andrej Karpathy's autonomous AI research agent ran 700 experiments in 2 days and gave a glimpse of where AI is heading
Reddit r/artificial

So cursor admits that Kimi K2.5 is the best open source model
Reddit r/LocalLLaMA