Ollama allows you to run open-source large language models,
such as gpt-oss, locally.
Ollama
bundles model weights, configuration, and data into a single package, defined by a Modelfile.
It optimizes setup and configuration details, including GPU usage.
For a complete list of supported models and model variants, see the Ollama model library.
Installation and Setup
Ollama installation
Follow these instructions to set up and run a local Ollama instance. Ollama will start as a background service automatically, if this is disabled, run:ollama pull <name-of-model>
to download a model from the Ollama model library:
- This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model.
- To view all pulled (downloaded) models, use
ollama list
langchain-ollama
partner package and run a model.
Ollama LangChain partner package install
Install the integration package with:LLM
Chat Models
Chat Ollama
Ollama tool calling
Ollama tool calling uses the OpenAI compatible web server specification, and can be used with the defaultBaseChatModel.bind_tools()
methods
as described here.
Make sure to select an ollama model that supports tool calling.