| 57M params, fully binary {-1,+1}, state space model. The C runtime doesn't include math.h — every operation is integer arithmetic (XNOR, popcount, int16 accumulator for SSM state). Designed for hardware without FPU: ESP32, Cortex-M, or anything with ~8MB of memory and a CPU. Also runs in browser via WASM. Trained on TinyStories so it generates children's stories — the point isn't competing with 7B models, it's running AI where nothing else can. [link] [comments] |
7MB binary-weight Mamba LLM — zero floating-point at inference, runs in browser
Reddit r/LocalLLaMA / 3/23/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- A linked Hugging Face space (OneBitModel/prisme) is provided as an example of the project’s accessibility and local/browser execution.
Related Articles

Composer 2: What is new and Compares with Claude Opus 4.6 & GPT-5.4
Dev.to
AI Text Analyzer vs Asking Friends: Which Gives Better Perspective?
Dev.to
[D] Cathie wood claims ai productivity wave is starting, data shows 43% of ceos save 8+ hours weekly
Reddit r/MachineLearning

Microsoft hires top AI researchers from Allen Institute for AI for Suleyman's Superintelligence team
THE DECODER
MolmoWeb 4B/8B
Reddit r/LocalLLaMA