Off Grid: Complete On-Device AI Suite for Mobile
Off Grid: The Swiss Army Knife of On-Device AI
Off Grid is not just another chat app—it’s a complete offline AI suite for your phone that respects your privacy. While most “local LLM” apps give you a text chatbot and call it a day, Off Grid delivers text generation, image generation, vision AI, voice transcription, and document analysis, all running natively on your phone’s hardware.
Zero data leaves your device. No cloud subscription. No data harvesting. Just powerful AI that works anywhere—even without an internet connection.
What Makes Off Grid Different
Most AI apps on phones are thin clients that send your data to remote servers. Off Grid flips this model entirely:
- ✅ 100% Offline — Everything runs on-device, no internet required
- ✅ No Account Needed — No signups, no email, no tracking
- ✅ Privacy by Design — Your conversations, photos, and documents never leave your phone
- ✅ Complete AI Suite — Not just chat: images, vision, voice, documents
- ✅ Open Source — MIT licensed, fully auditable code
- ✅ Bring Your Own Models — Use any GGUF model or Stable Diffusion checkpoint
Core Capabilities
💬 Text Generation
Run state-of-the-art language models directly on your phone:
Supported Models:
- Qwen 3 — Alibaba’s powerful multilingual model
- Llama 3.2 — Meta’s latest efficient architecture
- Gemma 3 — Google’s lightweight yet capable model
- Phi-4 — Microsoft’s compact reasoning specialist
- Any GGUF Model — Import your own
.gguffiles
Features:
- Streaming responses for real-time interaction
- Thinking mode for reasoning-heavy tasks
- 15-30 tokens/second on flagship devices
- Memory-efficient quantized models (4-bit, 5-bit, 8-bit)
Performance:
| Device | Speed | Model Size |
|---|---|---|
| Flagship (SD 8 Gen 3) | 15-30 tok/s | Up to 8B params |
| Mid-range | 5-15 tok/s | Up to 3B params |
🎨 Image Generation
Generate stunning images with on-device Stable Diffusion:
Key Features:
- Real-time Preview — See generation progress as it happens
- NPU Acceleration — Snapdragon devices generate images in 5-10 seconds
- Core ML Support — Optimized for Apple Silicon on iOS
- Multiple Models — 20+ checkpoints including:
- Absolute Reality (photorealistic)
- DreamShaper (artistic)
- Anything V5 (anime)
- Custom SD checkpoints
Performance:
| Device | NPU | CPU |
|---|---|---|
| Flagship | 5-10s | ~15s |
| Mid-range | — | ~30s |
AI Prompt Enhancement: Simple prompts get automatically expanded into detailed Stable Diffusion prompts by the text model, improving output quality without prompt engineering knowledge.
👁️ Vision AI
Point your camera at anything and ask questions:
Capabilities:
- Document Analysis — Read receipts, forms, contracts
- Scene Description — “What’s in this photo?”
- Object Recognition — Identify items, text, barcodes
- Visual Q&A — “How many calories are in this meal?”
Supported Models:
- SmolVLM — Fast, efficient vision model
- Qwen3-VL — Strong multilingual vision capabilities
- Gemma 3n — Google’s vision-enabled model
Performance:
- ~7 seconds on flagship devices
- ~15 seconds on mid-range phones
🎤 Voice Input
Transcribe speech to text without sending audio to servers:
Features:
- On-device Whisper — OpenAI’s speech recognition running locally
- Hold to Record — Press and hold, speak naturally, release to transcribe
- Auto-transcription — Text appears instantly when you stop recording
- Multi-language Support — Whisper handles 99 languages
Privacy Guarantee: No audio ever leaves your phone. Not even temporarily. Everything is processed locally in real-time.
📄 Document Analysis
Chat with your documents, completely offline:
Supported Formats:
- PDFs (native text extraction)
- Code files (Python, JavaScript, C++, etc.)
- CSV and data files
- Text documents
- Images (via vision AI)
Use Cases:
- Summarize long PDFs
- Extract key points from contracts
- Analyze code for bugs
- Query data from CSV files
- Compare multiple documents
Performance Benchmarks
Off Grid is optimized to run efficiently on mobile hardware:
| Task | Flagship | Mid-range |
|---|---|---|
| Text generation | 15-30 tok/s | 5-15 tok/s |
| Image gen (NPU) | 5-10s | — |
| Image gen (CPU) | ~15s | ~30s |
| Vision inference | ~7s | ~15s |
| Voice transcription | Real-time | Real-time |
Tested On:
- Snapdragon 8 Gen 2/3 (Android flagships)
- Apple A17 Pro (iPhone 15 Pro)
- Results vary by model size and quantization level
Installation
Android
Option 1: Download APK (Quickest)
- Visit GitHub Releases
- Download the latest APK
- Enable “Install from unknown sources” in Settings
- Install and start chatting in under 2 minutes
Option 2: Build from Source
git clone https://github.com/alichherawalla/off-grid-mobile.git
cd off-grid-mobile
npm install
# Android
cd android && ./gradlew clean && cd ..
npm run androidiOS
Build from Source:
git clone https://github.com/alichherawalla/off-grid-mobile.git
cd off-grid-mobile
npm install
# iOS
cd ios && pod install && cd ..
npm run iosRequirements:
- Node.js 20+
- JDK 17 / Android SDK 36 (Android)
- Xcode 15+ (iOS)
Use Cases
Privacy-Conscious Users
- Secure Conversations — Discuss sensitive topics without cloud logging
- Medical Queries — Ask health questions privately
- Financial Planning — Analyze documents without data exposure
- Journal Writing — Personal thoughts stay on-device
Travelers & Remote Workers
- Works Offline — AI assistant functions without internet
- No Roaming Charges — No data usage for AI tasks
- Rural Areas — Functional where connectivity is limited
- Airplane Mode — Use during flights
Developers & Tech Enthusiasts
- Experiment with Models — Test different LLMs and checkpoints
- Custom Workflows — Bring your own GGUF models
- Open Source — Modify, extend, contribute
- Learning — Understand on-device AI capabilities
Content Creators
- Generate Images — Create visuals without internet
- Voice Notes — Transcribe ideas instantly
- Document Drafting — Write offline, polish with AI
- Research — Analyze documents privately
Professionals
- Contract Review — Analyze legal documents offline
- Meeting Prep — Review materials privately
- Code Review — Check code without uploading
- Data Analysis — Query CSVs without cloud exposure
Technical Architecture
Off Grid leverages cutting-edge on-device AI frameworks:
Off Grid Mobile App
├── Text Generation (llama.cpp + llama.rn)
├── Image Generation (local-dream / ml-stable-diffusion)
├── Vision AI (MNN / Core ML)
├── Voice Transcription (whisper.cpp + whisper.rn)
└── Document Parsing (Native PDF / Text)Key Dependencies:
- llama.cpp — Efficient LLM inference
- whisper.cpp — Fast speech recognition
- llama.rn / whisper.rn — React Native bindings
- local-dream — On-device Stable Diffusion
- ml-stable-diffusion — Apple’s Core ML diffusion
- MNN — Alibaba’s lightweight inference engine
Privacy & Security
What Stays on Your Phone
✅ All conversations and chat history ✅ Generated images and documents ✅ Voice recordings and transcriptions ✅ Uploaded PDFs and files ✅ Model downloads and caches
What Leaves Your Phone
❌ Nothing. Zero data transmission.
Security Features
- No network permissions required for core functionality
- Optional model downloads over HTTPS
- Local storage encryption (OS-level)
- No analytics or telemetry
- No account creation = no personal data collection
Comparison with Cloud-Based AI Apps
| Feature | Off Grid | ChatGPT | Claude | Gemini |
|---|---|---|---|---|
| Offline Functionality | ✅ 100% | ❌ No | ❌ No | ❌ No |
| Privacy | ✅ Maximum | ⚠️ Cloud | ⚠️ Cloud | ⚠️ Cloud |
| No Subscription | ✅ Free | ⚠️ Limited | ❌ Paid | ⚠️ Limited |
| Image Generation | ✅ Built-in | ❌ Separate | ❌ No | ✅ Yes |
| Vision AI | ✅ Built-in | ✅ Yes | ✅ Yes | ✅ Yes |
| Voice Input | ✅ Offline | ⚠️ Cloud | ⚠️ Cloud | ⚠️ Cloud |
| Open Source | ✅ MIT | ❌ No | ❌ No | ❌ No |
| Custom Models | ✅ GGUF | ❌ No | ❌ No | ❌ No |
Choose Off Grid when: Privacy is paramount, you need offline functionality, or you want to avoid subscription fees.
Choose cloud apps when: You need the absolute latest models (GPT-4, Claude 3 Opus) or advanced features like web browsing.
Tips for Best Performance
Model Selection
- Start Small: Use 3B parameter models for quick responses
- Scale Up: Try 7-8B models for complex reasoning
- Quantization: Use Q4_K_M for balance of speed and quality
- Context Length: Shorter contexts run faster on mobile
Image Generation
- Use NPU: Snapdragon 8 Gen 2/3 devices get 3-5x speedup
- Lower Steps: 20-25 steps often sufficient for good quality
- Smaller Resolutions: 512x512 generates faster than 1024x1024
- Batch Size: Generate one image at a time on mobile
Battery & Thermals
- Long sessions: Take breaks to prevent overheating
- Charging: Performance better when plugged in
- Background: Close other apps for maximum RAM availability
- Model Management: Delete unused models to free storage
Storage Management
| Component | Size |
|---|---|
| 3B Parameter LLM | ~2 GB |
| 7B Parameter LLM | ~4-5 GB |
| Stable Diffusion Model | ~2-4 GB |
| Whisper Model | ~150 MB |
| Vision Model | ~1-2 GB |
Total: Plan for 10-20 GB depending on model selection.
Limitations & Considerations
- Model Size: Mobile hardware limits max model size (typically 8B parameters)
- Speed: Slower than cloud APIs (seconds vs. milliseconds)
- Battery Usage: AI inference is power-intensive
- Heat Generation: Extended use may warm your device
- Storage: Large models require significant space
- RAM: Minimum 6GB RAM recommended, 8GB+ ideal
References
- 🔗 GitHub Repository: https://github.com/alichherawalla/off-grid-mobile
- 🔗 Releases: GitHub Releases
- 🔗 Documentation: See repo docs/ folder for detailed guides
Related Technologies:
- llama.cpp — https://github.com/ggerganov/llama.cpp
- whisper.cpp — https://github.com/ggerganov/whisper.cpp
- Hugging Face — https://huggingface.co (for GGUF models)
- LocalAI — Desktop alternative for computers
Similar Mobile Apps:
- PocketPal — iOS on-device LLM chat
- ChatLLM — Android LLM interface
- MNN Chat — Alibaba’s mobile AI app
Why This Tool Rocks
- True Privacy: Not “privacy-focused”—actually private. Zero data transmission.
- Complete Independence: Works anywhere, anytime, regardless of connectivity
- No Gatekeeping: No accounts, no API keys, no subscriptions, no limits
- Full-Featured: Not a toy—production-capable text, image, vision, and voice AI
- Open Source: MIT license means free forever, community-improvable
- Customizable: Use any compatible model, not just what the vendor provides
- Fast Improving: Mobile NPU acceleration is getting better with each generation
- Future-Proof: As models get more efficient, performance improves automatically
In an era where AI companies want to own your data and rent you access, Off Grid represents a different vision: powerful AI that you truly own and control, running on hardware you already have.
Crepi il lupo! 🐺