setup
1. install Ollama & pull a model
ministral-3 is a good starting point (fast, works on most machines).
2. select Ollama in screenpipe
- open the screenpipe app
- click the AI preset selector (top of the chat/timeline)
- click Ollama
- pick your model from the dropdown (screenpipe auto-detects pulled models)
- start chatting
localhost:11434 automatically.
recommended models
| model | size | best for |
|---|---|---|
ministral-3 | ~2 GB | fast, general use, recommended starting point |
gemma3:4b | ~3 GB | strong quality for size, good for summaries |
qwen3:4b | ~3 GB | multilingual, good reasoning |
deepseek-r1:8b | ~5 GB | strong reasoning, needs 16 GB+ RAM |
requirements
- Ollama installed and running
- at least one model pulled
- screenpipe running
troubleshooting
“ollama not detected”- make sure Ollama is running:
ollama serve - check it’s responding:
curl http://localhost:11434/api/tags
- pull it first:
ollama pull ministral-3 - you can also type the model name manually in the input field
- try a smaller model (
ministral-3) - close other GPU-heavy apps
- ensure you have enough free RAM (model size + ~2 GB overhead)