We shipped Gemma 4 (E2B and E4B edge variants) in Off Grid today — our open-source, offline-first AI app for Android and iOS. What makes this different from other local LLM setups: → No server, no Python, no laptop. Runs entirely on your phone's NPU/CPU. → Gemma 4's 128K context window, fully on-device — finally useful for long docs and code on mobile. → Native vision: point your camera at anything and ask Gemma 4 about it. → Whisper speech-to-text, Stable Diffusion image gen, tool calling — all in one app. → ~15–30 tok/s on Snapdragon 8 Gen 3 / Apple A17 Pro. → Apache 2.0 model, MIT app — genuinely open all the way down. Gemma 4's E2B variant running in under 1.5GB RAM on a phone is honestly wild. The E4B with 128K context + vision is what we've been waiting for. Android (live now): https://play.google.com/store/apps/details?id=ai.offgridmobile iOS: coming soon GitHub (MIT): https://github.com/alichherawalla/off-grid-mobile-ai Would love to hear tok/s numbers people are seeing across different devices. Drop them below. [link] [comments]



