Running AI in Your Pocket: On-Device LLM App with MediaPipe

Your phone today is basically a tiny AI workstation. It can run a full language model offline, with zero cloud, and no data leaving your device. That was the whole point of PocketLLM, an Android app I built that runs Google’s Gemma 3n model locally using MediaPipe Tasks GenAI. No servers. No API keys (except initial HuggingFace model download). No monthly bills. Just pure on-device intelligence. Why I Wanted an On-Device LLM I’ve been experimenting with running AI locally for a while, first on my MacBook, then on Android. Some reasons why you could want this: ...

November 17, 2025 · 3 min