Skip to main content

Ollama

Run AI agents with local models using Ollama. No API key needed — everything runs on your machine.

Setup

  1. Install Ollama: ollama.com/download
  2. Pull a model:
ollama pull llama3.2
  1. Ollama starts automatically and serves on localhost:11434. Configure in droid.yaml:
ollama:
base_url: "http://localhost:11434/v1"

No API key required.

Agent Config

name: "local-agent"
model:
provider: "ollama"
name: "llama3.2"
max_tokens: 2048
temperature: 0.7
ModelSizeBest for
llama3.23BFast, general purpose
llama3.1:8b8BGood quality/speed balance
llama3.1:70b70BHigh quality (needs GPU)
mistral7BFast, multilingual
codellama7-34BCode generation
qwen2.57BMultilingual

See ollama.com/library for all available models.

Docker Networking

If Sinaptic® DROID+ runs in Docker and Ollama runs on the host:

ollama:
base_url: "http://host.docker.internal:11434/v1"

If both run in Docker, use Docker Compose networking:

services:
droid:
image: sinapticai/droid:latest
# ...
ollama:
image: ollama/ollama:latest
volumes:
- ollama-data:/root/.ollama

Then use http://ollama:11434/v1 as the base URL.

Notes

  • Ollama serves an OpenAI-compatible API, so it integrates seamlessly with Sinaptic® DROID+.
  • Tool use (function calling) support depends on the model. Llama 3.1+ and Mistral support function calling.
  • Performance depends entirely on your hardware. GPU is recommended for models larger than 7B.
  • Models are downloaded once and cached locally by Ollama.