Ollama#
Ollama lets you download and run open-source large language models locally. It handles model management, quantization, and exposes a local API that other tools (like OpenCode) can connect to.
Getting Started#
Download Ollama from ollama.com, install it, and pull a model:
ollama pull qwen3.5
ollama run qwen3.5See the Ollama documentation for the full list of available models and configuration options.
When to Use It#
Ollama is the simplest way to run models locally. Use it when you need offline access, want to keep data off external APIs, or want to experiment with different open-source models. See Running Models Locally for guidance on choosing models and hardware requirements.