Every other guide for accessing Ollama from your phone starts the same way: open a terminal, set environment variables, find your IP address, configure your firewall. By the time you are done, you have spent more time on setup than you will spend chatting with the model.
Your Android phone and your Ollama server are on the same network. They should just be able to talk to each other.
Off Grid makes that happen. It auto-discovers Ollama servers on your local network, pulls the model list, and lets you start chatting. No IP addresses, no port numbers, no configuration files on your phone.
What you need
- A computer running Ollama with at least one model
- An Android phone (6GB+ RAM recommended) on the same WiFi network
- Off Grid installed from GitHub Releases
Step 1: Open Ollama to your network
Ollama only listens on localhost by default. One change fixes that.
On Mac or Linux:
OLLAMA_HOST=0.0.0.0 ollama serve
To make it permanent, add export OLLAMA_HOST=0.0.0.0 to your .zshrc or .bashrc.
On Windows: Add OLLAMA_HOST as a system environment variable with value 0.0.0.0. Restart Ollama.
Done.
Step 2: Scan from Off Grid
Open Off Grid on your Android phone. Go to Remote Models. Tap Scan Network.
Off Grid finds your Ollama server, pulls the list of installed models, and shows them to you. If you have multiple machines running Ollama on your network, it finds all of them.
Tap a model. Start typing. Responses stream in.

Off Grid scanning the network and discovering Ollama models - iOS, Android, and servers running side by side.
Step 3: Use every model you have
The power of this setup is that you are no longer constrained by your phone's hardware. Your phone has 6-8GB of RAM and can run a 2B model. Your desktop might have 32GB+ of RAM and a GPU that can run 70B models comfortably.
Qwen 3.5 9B is the sweet spot for most setups. Released March 2026, it outperforms OpenAI's GPT-OSS-120B on reasoning and language benchmarks at 13 times smaller. If your computer has 16GB of RAM, it runs this model well. From your Android phone over WiFi, the experience is smooth and fast.
You can switch models at any time, even in the middle of a conversation. Ask a quick question with a small fast model, then switch to the 9B for a follow-up that needs more depth. The chat history stays intact.
What makes this more than a chat client
Other apps that connect to Ollama are just remote chat interfaces. Off Grid is a full AI toolkit.
Projects with RAG. Create a project, attach your documents - PDFs, code, CSVs, text files - and Off Grid builds a local knowledge base. When you ask a question, it searches your documents and feeds relevant context to the model. Your Ollama server does the inference, your phone manages the knowledge base. Everything stays private.
Tool calling. Models with function calling support (Qwen 3.5, Llama 3.1, Mistral) can use built-in tools: web search, calculator, date/time, device info. The model decides what tools to use and chains them together automatically.
On-device models too. Off Grid runs smaller models directly on your Android phone's hardware. Snapdragon 8 Gen 2 and newer phones get OpenCL GPU acceleration. This means you have AI even when your WiFi drops. The on-device model handles quick tasks, and the Ollama server handles the heavy lifting when you are home.
Vision, voice, documents. Point your camera at something and ask about it. Dictate with on-device Whisper transcription. Attach files to conversations. All of this works with both local and remote models.
You already have the hardware
You do not need to buy a dedicated server. You do not need a cloud GPU subscription. You do not need anything you do not already own.
Your desktop or laptop, the one you already use every day, can run Ollama in the background while you use it normally. The model inference uses your GPU when you are not gaming or rendering. Your phone connects over the WiFi you already pay for.
The cost of running private AI that rivals ChatGPT: the electricity to keep your computer on. That is it.
Where Off Grid is heading
We are building toward a personal AI operating system - all the compute you own, working together, completely private. Network discovery, on-device inference, projects, RAG, tool calling, vision, and voice are all here today. Automatic task routing, seamless device handoff, and shared context are next.
Want to be part of building this? Join the Off Grid Slack from our GitHub.
Try it
- GitHub (1,000+ stars, 10,000+ downloads in 4 weeks)
- Android: GitHub Releases
- iOS: App Store (search "Off Grid AI")
Your Ollama server is running. Your phone is in your pocket. Off Grid connects them.
Off Grid is built by the team at Wednesday Solutions, a product engineering company with a 4.8/5.0 rating on Clutch across 23 reviews.

Top comments (0)