Author: Julian A. Gonzalez, IBM Champion 2025
Date: 6-29-2025
Overview: A beginner-friendly guide to running IBMβs Granite 3.2-2B language model on Raspberry Pi 5 using only built-in hardware. Create a local AI assistant with CPU-only processing.
| Component | Minimum | Recommended | |ββββ|βββ|ββββ-| | Raspberry Pi | Pi 5 (4GB) | Pi 5 (8GB) | | Storage | 64GB Class 10 microSD | Higher speed card | | Power Supply | Standard USB-C | 27W Official PSU | | Cooling | Heatsink | Active cooling case | | Network | Wi-Fi | Ethernet connection |
sudo apt update && sudo apt upgrade -y
curl -fsSL https://ollama.com/install.sh | sh
ollama pull granite3.2:2b
# Alternative custom setup if needed
ollama create granite-pi -f granite-pi.Modelfile
Start interactive session:
ollama run granite-pi
Sample prompts:
/bye
or Ctrl+C
| Pi Model | Response Time | Best Use Cases | |βββ-|βββββ|βββββ| | Pi 5 (4GB) | 10-20 seconds | Simple questions | | Pi 5 (8GB) | 5-15 seconds | Longer conversations |
| Issue | Solution | |ββ-|βββ-| | Slow responses | Check cooling, verify performance mode | | Out of memory | Close apps, reboot, reduce context size | | System freezes | Improve cooling, use 8GB model | | Model wonβt load | Restart Ollama, check memory |
~/chat.sh
Note: The Pi will get warm during operation - ensure proper cooling!
Explore the possibilities of local AI on affordable hardware!
Full Guide Details | Ollama GitHub | IBM Granite Docs |