Basic PSA. PocketPal got updated, so runs Gemma 4.

Reddit r/LocalLLaMA / 4/5/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • PocketPal was updated recently and can now run Meta’s Gemma 4 models (2B and 4B) well on the author’s Android device.
  • The post highlights that smaller parameter models run quickly and reliably under PocketPal, making them more feasible for low-RAM phones.
  • The author plans to attempt running a much larger Gemma 26B quantization on a 12GB RAM phone but expects Android overhead to prevent it.
  • The update is framed as part of an ongoing trend toward more capable local/edge model execution, with future targets like 7–8B-sized models.
  • A GitHub link to PocketPal is provided for users who want to get the update.

Just because I've seen a couple of "I want this on Android" questions, PocketPal got updated a few hours ago, and runs Gemma 4 2B and 4B fine. At least on my hardware (crappy little moto g84 workhorse phone). Love an app that gets regular updates.

I'm going to try and squeak 26B a4 iq2 quantization into 12gigs of ram, on a fresh boot, but I'm almost certain it can't be done due to Android bloat.

But yeah, 2B and 4B work fine and quickly under PocketPal. Hopefully their next one is 7-8B (not 9B), because the new Qwen 3.5 models just skip over memory caps, but the old ones didn't. Super numbers are great, running them with OS overhead and context size needs a bit smaller, to be functional on a 12gig RAM phone.

Bring on the GemmaSutra 4 4B though, as another gold standard of thinking's and quick ish. We will fix her. We have the technology!

https://github.com/a-ghorbani/pocketpal-ai

submitted by /u/Sambojin1
[link] [comments]