Ollama Setup
Ollama Setup
Run AI locally with Ollama for privacy and offline usage.
Install Ollama
macOS
Linux
Windows
Download from ollama.com
Start Ollama
This starts the Ollama server on localhost:11434.
Pull a Model
Usage
CLI
Config File
Programmatic
Recommended Models
| Model | Size | Quality | Speed |
|---|---|---|---|
llama3.2 |
3B | Good | Fast |
llama3.2:7b |
7B | Better | Medium |
mistral |
7B | Good | Medium |
gemma2 |
9B | Good | Medium |
Recommended: llama3.2 for balance of speed and quality.
Features
All AI features work with Ollama:
- Meta description generation
- Alt text generation (limited)
- Title suggestions
- Content analysis
Note: Image analysis (alt text) may be limited depending on model.
No API Key Needed
Ollama runs locally, no API key required:
Advantages
- Privacy - Data never leaves your machine
- Offline - Works without internet
- Free - No API costs
- Control - Choose any model
Disadvantages
- Quality - May be lower than cloud APIs
- Resources - Requires local GPU/CPU
- Setup - More complex than cloud
- Speed - Depends on hardware
System Requirements
| Model Size | RAM | GPU (Optional) |
|---|---|---|
| 3B | 4GB | 4GB VRAM |
| 7B | 8GB | 8GB VRAM |
| 13B | 16GB | 16GB VRAM |
Troubleshooting
"Connection refused"
Ollama server not running:
"Model not found"
Pull the model first:
Slow responses
- Use smaller model (
llama3.2vsllama3.2:7b) - Ensure GPU is being used:
Out of memory
- Use smaller model
- Close other applications
- Increase swap space
Check Ollama Status
Comparison
| Feature | Ollama | Cloud APIs |
|---|---|---|
| Cost | Free | Pay per use |
| Privacy | Full | Data sent to API |
| Offline | Yes | No |
| Quality | Variable | Consistent |
| Setup | Complex | Simple |
| Speed | Hardware dependent | Fast |
Use Ollama when:
- Privacy is critical
- Offline access needed
- Cost is a concern
- You have capable hardware