Ollama
Install Ollama
Download and install Ollama from ollama.ai.
Verify the local server
Ollama runs a local server at
localhost:11434. It starts automatically after installation.LM Studio
Install LM Studio
Download and install LM Studio from lmstudio.ai.
Hardware Recommendations
Apple Silicon (M1 or later) with 16GB+ RAM is recommended for local models. Larger models (13B+ parameters) benefit from 32GB+ RAM. Apple’s unified memory architecture makes M-series chips particularly well-suited for local inference.
| Model Size | Minimum RAM | Recommended RAM |
|---|---|---|
| 7B parameters | 8 GB | 16 GB |
| 13B parameters | 16 GB | 32 GB |
| 34B+ parameters | 32 GB | 64 GB |
Privacy
With local models, your code never touches any network. Every request is processed entirely on your Mac. This is ideal for:- Proprietary codebases where source code cannot leave the organization
- Regulated industries with strict data residency requirements
- Air-gapped environments with no internet access
- Any scenario where you need complete control over where your data goes
Local models trade some capability for privacy. Cloud-hosted models like Claude Opus 4.5 and GPT-5.2 generally produce higher-quality results for complex tasks. Choose based on your privacy requirements and the complexity of your work.
