Overview
Integration details
Class | Package | Local | Serializable | PY support | Downloads | Version |
---|---|---|---|---|---|---|
AzureChatOpenAI | @langchain/openai | ❌ | ✅ | ✅ |
Model features
See the links in the table headers below for guides on how to use specific features.Tool calling | Structured output | JSON mode | Image input | Audio input | Video input | Token-level streaming | Token usage | Logprobs |
---|---|---|---|---|---|---|---|---|
✅ | ✅ | ✅ | ✅ | ❌ | ❌ | ✅ | ✅ | ✅ |
Setup
Azure OpenAI is a cloud service to help you quickly develop generative AI experiences with a diverse set of prebuilt and curated models from OpenAI, Meta and beyond. LangChain.js supports integration with Azure OpenAI using the new Azure integration in the OpenAI SDK. You can learn more about Azure OpenAI and its difference with the OpenAI API on this page.Credentials
If you don’t have an Azure account, you can create a free account to get started. You’ll also need to have an Azure OpenAI instance deployed. You can deploy a version on Azure Portal following this guide. Once you have your instance running, make sure you have the name of your instance and key. You can find the key in the Azure Portal, under the “Keys and Endpoint” section of your instance. Then, if using Node.js, you can set your credentials as environment variables:Installation
The LangChain AzureChatOpenAI integration lives in the@langchain/openai
package:
Instantiation
Now we can instantiate our model object and generate chat completions:Invocation
Chaining
We can chain our model with a prompt template like so:Using Azure Managed Identity
If you’re using Azure Managed Identity, you can configure the credentials like this:Using a different domain
If your instance is hosted under a domain other than the defaultopenai.azure.com
, you’ll need to use the alternate AZURE_OPENAI_BASE_PATH
environment variable.
For example, here’s how you would connect to the domain https://westeurope.api.microsoft.com/openai/deployments/{DEPLOYMENT_NAME}
:
Custom headers
You can specify custom headers by passing in aconfiguration
field:
configuration
field also accepts other ClientOptions
parameters accepted by the official SDK.
Note: The specific header api-key
currently cannot be overridden in this manner and will pass through the value from azureOpenAIApiKey
.
Migration from Azure OpenAI SDK
If you are using the deprecated Azure OpenAI SDK with the@langchain/azure-openai
package, you can update your code to use the new Azure integration following these steps:
- Install the new
@langchain/openai
package and remove the previous@langchain/azure-openai
package:
-
Update your imports to use the new
AzureChatOpenAI
class from the@langchain/openai
package: -
Update your code to use the new
AzureChatOpenAI
class and pass the required parameters:Notice that the constructor now requires theazureOpenAIApiInstanceName
parameter instead of theazureOpenAIEndpoint
parameter, and adds theazureOpenAIApiVersion
parameter to specify the API version.-
If you were using Azure Managed Identity, you now need to use the
azureADTokenProvider
parameter to the constructor instead ofcredentials
, see the Azure Managed Identity section for more details. -
If you were using environment variables, you now have to set the
AZURE_OPENAI_API_INSTANCE_NAME
environment variable instead ofAZURE_OPENAI_API_ENDPOINT
, and add theAZURE_OPENAI_API_VERSION
environment variable to specify the API version.
-
If you were using Azure Managed Identity, you now need to use the