Ollama
Trace Ollama local LLM calls with Respan.
Set up Respan
- Sign up — Create an account at platform.respan.ai
- Create an API key — Generate one on the API keys page
- Add credits or a provider key — Add credits on the Credits page or connect your own provider key on the Integrations page
Use AI
Add the Docs MCP to your AI coding tool to get help building with Respan. No API key needed.
What is Ollama?
Ollama lets you run large language models locally. It provides a simple CLI and API for downloading, running, and managing models like Llama, Mistral, and Gemma on your own hardware.
Setup
What gets traced
- Model name and version
- Prompt and completion tokens
- Input/output content
- Response latency
- Streaming chunks
Traces appear in the Traces dashboard.
Learn more
- OpenTelemetry integration - How Respan processes OTel spans
- Respan Python SDK - Full SDK reference