Run 70B+ AI models on your Mac and chat from your iPhone — everything stays on your network.
OwnPodAI Local Server works with any computer running Ollama — MacBook, Mac Mini, Mac Studio, Linux PC, or Windows PC. The only requirement: same WiFi network as your iPhone.
| Device Example | RAM | Chip | Models | Quality |
|---|---|---|---|---|
| MacBook Air M1 | 8 GB | M1 | Llama 3.2 3B, Phi-4 Mini, Qwen 3 4B | Basic |
| MacBook Air M2/M3 | 16 GB | M2/M3 | Llama 3.2 8B, Gemma 2 9B, Qwen 3 8B, Mistral 7B | Good |
| MacBook Pro M3/M4 | 18 GB | M3/M4 | Same as 16GB + slightly faster generation | Good |
| MacBook Pro M3 Pro/M4 Pro | 36 GB | M3/M4 Pro | Qwen 2.5 14B, Gemma 2 27B, CodeLlama 13B | Great |
| Mac Mini M4 | 32 GB | M4 | Llama 13B, Qwen 14B, Gemma 27B, Mistral 22B | Great |
| Mac Mini M4 Pro | 48 GB | M4 Pro | Llama 3.1 70B (Q3), Qwen 3 32B, Command R 35B | Excellent |
| Mac Mini M4 Pro / Mac Studio | 64 GB | M4 Pro/Max | Llama 3.1 70B, Qwen 3 72B, DeepSeek V3, Mixtral 8x7B | Excellent |
| Mac Studio M2 Ultra | 128 GB | M2 Ultra | Llama 3.1 405B (Q4), any model at high quality | Maximum |
| Mac Pro / Mac Studio M4 Ultra | 192 GB | M4 Ultra | Every model at full precision, multiple models simultaneously | Maximum |
Open Terminal and run:
# Install Ollama (one command) curl -fsSL https://ollama.com/install.sh | sh
Or download from ollama.com — drag to Applications, done.
Verify installation:
ollama --version # Should show: ollama version 0.x.x
curl -fsSL https://ollama.com/install.sh | sh
For NVIDIA GPU support, make sure CUDA drivers are installed first.
Download the installer from ollama.com/download/windows and run it. Ollama runs inside WSL2 automatically.
Pull any model with one command. Here are the best options:
# Recommended for 16GB Mac ollama pull llama3.2 # Best for 32GB Mac ollama pull qwen2.5:14b # Best for 64GB Mac (GPT-4 quality) ollama pull llama3.1:70b # Qwen 3 72B (best open model as of 2025) ollama pull qwen3:72b # Vision model — can understand images ollama pull llava:13b
llama3.2 to test your setup. It downloads fast and runs on any Mac. Upgrade to bigger models once everything works.By default, Ollama only listens on localhost. To access it from your iPhone, start it with network access:
# Allow connections from your local network OLLAMA_HOST=0.0.0.0 ollama serve
Keep this Terminal window open. Ollama is now listening on port 11434.
To auto-start Ollama with network access every time your Mac boots:
# Add to your shell profile echo 'export OLLAMA_HOST=0.0.0.0' >> ~/.zshrc source ~/.zshrc
You need your Mac's local IP to connect from the iPhone:
# Get your local IP ipconfig getifaddr en0 # Example output: 192.168.1.XXX
Or go to System Settings → WiFi → Details → IP Address.
Verify Ollama is accessible from your network:
# Test from your Mac (should return "Ollama is running") curl http://192.168.1.XXX:11434 # Test a model response curl http://192.168.1.XXX:11434/api/generate -d '{"model":"llama3.2","prompt":"Hello"}'
On your iPhone:
192.168.1.XXX11434OLLAMA_HOST=0.0.0.0http://192.168.1.XXX:11434 in Safariollama pull llama3.1:70b-q4_0ollama ps to see loaded modelsollama stop modelname# List downloaded models ollama list # See running models and memory usage ollama ps # Remove a model to free space ollama rm llama3.1:70b # Update a model to latest version ollama pull llama3.2 # Stop all running models ollama stop llama3.2 # Run a quick test in terminal ollama run llama3.2 "What is 2+2?" # Check Ollama version ollama --version
OLLAMA_HOST=0.0.0.0 setting is safe on home WiFi but should never be used on public networks or with port forwarding.