Skip to main content
Ollama lets you run AI models locally on your machine. screenpipe integrates natively with Ollama — no API keys, no cloud, completely private.

setup

1. install Ollama & pull a model

# install from https://ollama.com then:
ollama run ministral-3
this downloads the model and starts Ollama. you can use any model — ministral-3 is a good starting point (fast, works on most machines).

2. select Ollama in screenpipe

  1. open the screenpipe app
  2. click the AI preset selector (top of the chat/timeline)
  3. click Ollama
  4. pick your model from the dropdown (screenpipe auto-detects pulled models)
  5. start chatting
that’s it. screenpipe talks to Ollama on localhost:11434 automatically.
modelsizebest for
ministral-3~2 GBfast, general use, recommended starting point
gemma3:4b~3 GBstrong quality for size, good for summaries
qwen3:4b~3 GBmultilingual, good reasoning
deepseek-r1:8b~5 GBstrong reasoning, needs 16 GB+ RAM
pull any model with:
ollama pull <model-name>

requirements

  • Ollama installed and running
  • at least one model pulled
  • screenpipe running

troubleshooting

“ollama not detected”
  • make sure Ollama is running: ollama serve
  • check it’s responding: curl http://localhost:11434/api/tags
model not showing in dropdown?
  • pull it first: ollama pull ministral-3
  • you can also type the model name manually in the input field
slow responses?
  • try a smaller model (ministral-3)
  • close other GPU-heavy apps
  • ensure you have enough free RAM (model size + ~2 GB overhead)