We just open-sourced MOSS-TTS-Nano, a tiny multilingual speech generation model from MOSI.AI and the OpenMOSS team.
Some highlights:
- 0.1B parameters
- Realtime speech generation
- Runs on CPU without requiring a GPU
- Multilingual support (Chinese, English, Japanese, Korean, Arabic, and more)
- Streaming inference
- Long-text voice cloning
- Simple local deployment with
infer.py,app.py, and CLI commands
The project is aimed at practical TTS deployment: small footprint, low latency, and easy local setup for demos, lightweight services, and product integration.
GitHub:
https://github.com/OpenMOSS/MOSS-TTS-Nano
Huggingface:
https://huggingface.co/spaces/OpenMOSS-Team/MOSS-TTS-Nano
Online demo:
https://openmoss.github.io/MOSS-TTS-Nano-Demo/
Would love to hear feedback on quality, latency, and what use cases you’d want to try with a tiny open TTS model.
[link] [comments]




