LIVE TAPE
OpenClaw 88,412 stars·CVE-2026-25898 disclosed (HIGH, Hermes)·Hermes Agent v2026.4.7 published·Hermes Agent +182 stars (last hour)·OpenClaw v2026.4.6 — credential vault hardening·CVE-2026-26133 patched (NanoClaw)·Pi 5 16GB rumoured for Q3 — recheck guidance·Nanobot +47 stars (last hour)·ZeroClaw v0.4.2 — Apple container fixes·Mac Mini M4 wins quarterly hardware survey·OpenClaw 88,412 stars·CVE-2026-25898 disclosed (HIGH, Hermes)·Hermes Agent v2026.4.7 published·Hermes Agent +182 stars (last hour)·OpenClaw v2026.4.6 — credential vault hardening·CVE-2026-26133 patched (NanoClaw)·Pi 5 16GB rumoured for Q3 — recheck guidance·Nanobot +47 stars (last hour)·ZeroClaw v0.4.2 — Apple container fixes·Mac Mini M4 wins quarterly hardware survey·
PocketClawvol. 1 · 2026

Ollama + Phi-3 mini on a Raspberry Pi 5

Install Ollama and the smallest credible local LLM (Phi-3 mini 3.8B Q4) on a Raspberry Pi 5. Useful as fallback or for narrow tasks.

Prerequisites

  • Raspberry Pi 5 (8 GB recommended)
  • Active cooler installed
  • At least 4 GB free disk for the model

Steps

  1. Install Ollama

    Ollama has ARM64 builds that work on the Pi 5. Use the convenience script.

    curl -fsSL https://ollama.com/install.sh | sh
  2. Pull Phi-3 mini

    The 3.8B parameter Q4 quantised model. ~2.3 GB download. Larger models won't run usably on the Pi 5 — don't try Llama 3 8B here unless you want to be disappointed.

    ollama pull phi3:mini
  3. Test it

    Run a simple prompt to verify.

    ollama run phi3:mini "Write a one-sentence summary of the OpenClaw security crisis."
  4. Verify performance

    Expect roughly 6 tokens per second on the Pi 5. If you're getting under 4, check thermals — the active cooler should keep the Pi 5 below 70°C under sustained load.

    vcgencmd measure_temp
  5. Expose Ollama to other tools

    Ollama exposes an OpenAI-compatible API on port 11434 by default. Tools like Hermes Agent, Nanobot or your own scripts can call it as if it were OpenAI.

    # On the Pi:
    curl http://localhost:11434/api/generate \
      -d '{"model": "phi3:mini", "prompt": "Hello"}'
    
    # To allow other devices on the network to call it:
    sudo systemctl edit ollama.service
    # Add: Environment="OLLAMA_HOST=0.0.0.0"
    sudo systemctl daemon-reload
    sudo systemctl restart ollama

Troubleshooting

Inference very slow (under 3 tok/s)
Check thermal throttling: `vcgencmd get_throttled` should return `throttled=0x0`. If not, improve cooling. Also ensure no other heavy processes are running.
Out of memory crash
Don't try larger models on the Pi 5. Phi-3 mini Q4 is the realistic ceiling. If you need 7B class models, you need a mini PC or Mac Mini.

Where to go from here

Wire Ollama into Hermes Agent or Nanobot as a fallback LLM provider. Configure the agent to call cloud Claude for hard tasks and local Phi-3 mini for cheap subtasks.

Other tutorials
intermediate
Hermes Agent on a Raspberry Pi 5
End-to-end install of Hermes Agent on a fresh Raspberry Pi 5 (8 GB), accessed via Tailscale, with Cl…
beginner
Tailscale for self-hosted AI dashboards
Set up Tailscale to access your agent dashboard from anywhere without exposing it on the public inte…
beginner
Caddy reverse proxy with HTTPS for a self-hosted AI dashboard
Front your agent dashboard with Caddy on port 443 with automatic HTTPS via Let's Encrypt — no certbo…
intermediate
Migrate OpenClaw to Hermes Agent — step by step
Concrete migration path with shell commands. Snapshot OpenClaw, install Hermes, port tools, migrate …