> ## Documentation Index
> Fetch the complete documentation index at: https://docs.langchain.com/llms.txt
> Use this file to discover all available pages before exploring further.

# Model providers

The [Playground](/langsmith/prompt-engineering-concepts#playground) supports a wide range of model providers. You can select a provider, configure your preferred settings, and save these configurations to reuse across multiple prompts.

Use this page for a list of the available providers and their configuration options:

<div className="grid grid-cols-1 md:grid-cols-3 gap-3">
  <a href="#amazon-bedrock" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/bedrock.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=d6fc326f6c429075add5a742c87d12b1" alt="" width="65" height="65" data-path="images/providers/light/bedrock.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/bedrock.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=dc593829dc9ff5ad061f991e08bbebc8" alt="" width="65" height="65" data-path="images/providers/dark/bedrock.svg" />

    <span className="font-semibold">Amazon Bedrock</span>
  </a>

  <a href="#anthropic" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/anthropic.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=f681e2a9025c2e1104d4260c6b27d638" alt="" width="65" height="65" data-path="images/providers/light/anthropic.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/anthropic.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=02fc4a5cd4c124dbec46869777a15f34" alt="" width="65" height="65" data-path="images/providers/dark/anthropic.svg" />

    <span className="font-semibold">Anthropic</span>
  </a>

  <a href="#azure-openai" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/microsoft.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=9189910834f230f7f94115d88e7ec4bf" alt="" width="65" height="65" data-path="images/providers/light/microsoft.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/microsoft.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=36998a752a161152e010bc2c9b421567" alt="" width="65" height="65" data-path="images/providers/dark/microsoft.svg" />

    <span className="font-semibold">Azure OpenAI</span>
  </a>

  <a href="#deepseek" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/deepseek.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=5a11e956e2ba24fd746aebd929ee1a5e" alt="" width="65" height="65" data-path="images/providers/light/deepseek.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/deepseek.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=0cb62f13b3b37e89c35060913ca67515" alt="" width="65" height="65" data-path="images/providers/dark/deepseek.svg" />

    <span className="font-semibold">DeepSeek</span>
  </a>

  <a href="#fireworks" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/fireworks.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=a0b3bbc0a57a2753099203eb71ed191b" alt="" width="65" height="65" data-path="images/providers/light/fireworks.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/fireworks.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=b8d8b361e57f9f8547170901dd7df346" alt="" width="65" height="65" data-path="images/providers/dark/fireworks.svg" />

    <span className="font-semibold">Fireworks</span>
  </a>

  <a href="#google-gemini" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/gemini.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=486f57c43a26c9661700c37d0f575c05" alt="" width="65" height="65" data-path="images/providers/light/gemini.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/gemini.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=6436cb2ebecd9119e486f81fb2fdd360" alt="" width="65" height="65" data-path="images/providers/dark/gemini.svg" />

    <span className="font-semibold">Google Gemini</span>
  </a>

  <a href="#google-vertex-ai" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/gemini.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=486f57c43a26c9661700c37d0f575c05" alt="" width="65" height="65" data-path="images/providers/light/gemini.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/gemini.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=6436cb2ebecd9119e486f81fb2fdd360" alt="" width="65" height="65" data-path="images/providers/dark/gemini.svg" />

    <span className="font-semibold">Google Vertex AI</span>
  </a>

  <a href="#groq" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/groq.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=3f3d47f8a0619a4b2a81ecf11743caf1" alt="" width="65" height="65" data-path="images/providers/light/groq.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/groq.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=14b1b2fbe2756bf8b407684f33ffa722" alt="" width="65" height="65" data-path="images/providers/dark/groq.svg" />

    <span className="font-semibold">Groq</span>
  </a>

  <a href="#mistral-ai" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/mistral.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=2b060bfbf269ab1b7f41f3a835f7a0b6" alt="" width="65" height="65" data-path="images/providers/light/mistral.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/mistral.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=59bde8ab2250b2ba2deaa84b74a66e1b" alt="" width="65" height="65" data-path="images/providers/dark/mistral.svg" />

    <span className="font-semibold">Mistral AI</span>
  </a>

  <a href="#openai" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/light/openai.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=fb29ba01edf0644930f3eda68ab6c0c2" alt="" width="65" height="67" data-path="images/providers/light/openai.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/openai.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=ad31e0e26cb456a8fa3cddea3bac876e" alt="" width="65" height="67" data-path="images/providers/dark/openai.svg" />

    <span className="font-semibold">OpenAI</span>
  </a>

  <a href="#openai-compatible-endpoint" className="flex items-center justify-center gap-2 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <Icon icon="link" iconType="solid" className="w-5 h-5 flex-shrink-0 mx-1" />

    <span className="font-semibold">OpenAI compatible endpoint</span>
  </a>

  <a href="#xai" className="flex items-center justify-center gap-1.5 p-2 rounded-lg border border-gray-200 dark:border-gray-700 hover:border-gray-300 dark:hover:border-gray-600 no-underline">
    <img className="block dark:hidden w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/z7oQGiHwXv52HwOy/images/providers/light/xai.svg?fit=max&auto=format&n=z7oQGiHwXv52HwOy&q=85&s=9e2e1244434658c80b07b419d4af6db9" alt="" width="65" height="65" data-path="images/providers/light/xai.svg" />

    <img className="hidden dark:block w-5 h-5" src="https://mintcdn.com/langchain-5e9cc07a/ZPKed1feKJ8F6LVo/images/providers/dark/xai.svg?fit=max&auto=format&n=ZPKed1feKJ8F6LVo&q=85&s=4d6a343007a7831ce03cfef3e78eebc6" alt="" width="65" height="65" data-path="images/providers/dark/xai.svg" />

    <span className="font-semibold">XAI</span>
  </a>
</div>

For details on creating and managing model configurations, refer to the [Configure prompt settings](/langsmith/managing-model-configurations) page.

## Amazon Bedrock

Before you use this model, ensure you have [AWS credentials or IAM role](https://docs.aws.amazon.com/bedrock/latest/userguide/security-iam.html).

### Authentication

Amazon Bedrock supports two authentication methods. **IAM trusted entity is the recommended approach** because it avoids sharing long-lived AWS access keys with LangSmith.

#### IAM trusted entity (recommended)

With IAM trusted entity authentication, you create an IAM role in your AWS account and allow LangSmith to assume it. No access keys are stored in LangSmith. Instead, LangSmith uses [AWS STS](https://docs.aws.amazon.com/STS/latest/APIReference/welcome.html) to assume the role on each request.

To set this up:

1. Create an IAM role in your AWS account with permissions to invoke Bedrock models (e.g., `bedrock:InvokeModel`).
2. Add a trust policy that allows LangSmith's AWS account (`808407022534`) to assume the role, using your LangSmith workspace ID as the external ID:

```json theme={"theme":{"light":"catppuccin-latte","dark":"catppuccin-mocha"}}
{
  "Version": "2012-10-17",
  "Statement": [
    {
      "Effect": "Allow",
      "Principal": {
        "AWS": "arn:aws:iam::808407022534:root"
      },
      "Action": "sts:AssumeRole",
      "Condition": {
        "StringEquals": {
          "sts:ExternalId": "<your-langsmith-workspace-id>"
        }
      }
    }
  ]
}
```

<Tip>
  You can find your workspace ID in your [LangSmith workspace settings](https://smith.langchain.com/settings).
</Tip>

3. In the LangSmith Playground, expand the **IAM Trusted Entity** section under the Bedrock provider and enter the ARN of the role you created.

For more details on trust policies, see the [AWS documentation](https://aws.amazon.com/blogs/security/how-to-use-trust-policies-with-iam-roles/).

#### Access keys

Alternatively, you can authenticate with AWS access keys (`AWS_ACCESS_KEY_ID` and `AWS_SECRET_ACCESS_KEY`). Enter these in the Bedrock provider configuration in the Playground. This method is simpler to set up but less secure because it requires storing long-lived credentials.

### Available models

AWS Bedrock provides access to foundation models from multiple providers:

* **Anthropic:** Claude models.
* **Amazon:** Titan models.
* **Cohere:** Command models.
* **Meta:** Llama models.
* **Others:** Additional providers available based on region.

For the current list of available models, refer to the [AWS Bedrock documentation](https://docs.aws.amazon.com/bedrock/latest/userguide/models-supported.html).

### Configuration parameters

Parameters depend on the underlying model provider:

#### For Anthropic models

Uses Anthropic configuration (see [Anthropic](#anthropic) section below).

#### For Amazon Titan

| Parameter       | Range     | Description             |
| --------------- | --------- | ----------------------- |
| **Temperature** | 0.0 - 1.0 | Response randomness     |
| **Max Tokens**  | 1+        | Maximum response length |
| **Top P**       | 0.0 - 1.0 | Nucleus sampling        |

#### AWS-specific settings

* **Region:** AWS region for model deployment.

### Tool calling

Depends on underlying model:

* **Anthropic models:** `auto`, `any`.
* **Cohere models:** `auto`.

## Anthropic

Before you use this model, ensure you have an [Anthropic API key](https://console.anthropic.com/settings/keys).

### Available models

Anthropic offers three tiers of models across their Claude generations:

* **Opus:** Highest intelligence and capability.
* **Sonnet:** Balanced performance and cost.
* **Haiku:** Fast and cost-effective.

Recent Claude models support extended thinking capabilities for showing reasoning processes.

For the current list of available models, refer to the [Anthropic documentation](https://docs.anthropic.com/claude/docs/models-overview).

### Configuration parameters

| Parameter             | Range     | Default  | Description                                        |
| --------------------- | --------- | -------- | -------------------------------------------------- |
| **Temperature**       | 0.0 - 1.0 | Optional | Randomness control (uncheck to use model default)  |
| **Max Output Tokens** | 1+        | 1024     | Maximum response length                            |
| **Top P**             | 0.0 - 1.0 | Optional | Nucleus sampling (uncheck for model default)       |
| **Top K**             | 1+        | Optional | Limits to top K tokens (uncheck for model default) |

<Note>
  Temperature, Top P, and Top K are optional. When unchecked, Claude uses its internal defaults.
</Note>

#### Extended Thinking

Available on supported Claude models. Enable the model to show reasoning before responding, similar to OpenAI's o-series.

| Parameter                    | Range  | Description                             |
| ---------------------------- | ------ | --------------------------------------- |
| **Enable Extended Thinking** | Toggle | Show/hide thinking process              |
| **Budget Tokens**            | 1+     | Max tokens for thinking (default: 1024) |

When enabled, responses include:

1. A "thinking" section with the model's reasoning.
2. The final response.

#### Advanced options

* **Base URL:** Override API endpoint for custom deployments.

### Tool calling

* **Supported Tool Choices:** `auto`, `any` (requires at least one tool).
* **Parallel Execution:** No (sequential only).

## Azure OpenAI

Before you use this model, ensure you have [Azure OpenAI credentials](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart) (endpoint + API key).

### Available models

Azure OpenAI provides the same model families as OpenAI:

* **GPT series:** General-purpose chat models.
* **o-series:** Reasoning-focused models.
* **Legacy models:** GPT-3.5 and GPT-4 variants.

Model availability varies by Azure region and requires deployment before use.

For the current list of available models, refer to the [Azure OpenAI documentation](https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models).

### Configuration parameters

Azure OpenAI supports the same parameters as OpenAI:

#### Standard parameters

| Parameter             | Range      | Description                                                        |
| --------------------- | ---------- | ------------------------------------------------------------------ |
| **Temperature**       | 0.0 - 2.0  | Controls randomness. Lower = more focused, higher = more creative. |
| **Max Output Tokens** | 1+         | Maximum length of the response                                     |
| **Top P**             | 0.0 - 1.0  | Nucleus sampling threshold. Alternative to temperature.            |
| **Presence Penalty**  | -2.0 - 2.0 | Penalize new topics (positive) or encourage them (negative)        |
| **Frequency Penalty** | -2.0 - 2.0 | Penalize repetition (positive) or allow it (negative)              |
| **Seed**              | Integer    | For reproducible outputs                                           |

#### Advanced parameters

**Reasoning Effort:** Available on reasoning-optimized models (o-series and newer GPT models).

**Service Tier:** Available on newer models.

**Other parameters:**

* **JSON Mode:** Force valid JSON responses.
* **Parallel Tool Calls:** Execute multiple tools concurrently.

#### Azure-specific features

* **Deployment Management:** Models must be deployed before use.
* **Regional Availability:** Choose Azure regions for data residency.
* **Content Filtering:** Built-in content moderation and safety features.
* **Managed Identity:** Azure AD authentication support.
* **Private Endpoints:** VNet integration for secure access.

### Tool calling

* **Supported Tool Choices:** `auto`, `required`, `none`, or specific tool name.
* **Parallel Execution:** Yes.

## DeepSeek

Before you use this model, ensure you have a [DeepSeek API key](https://platform.deepseek.com/api_keys).

### Available models

DeepSeek offers general-purpose models, reasoning-optimized models (R-series), and coding-specialized models.

For the current list of available models, refer to [DeepSeek's documentation](https://platform.deepseek.com/api-docs/).

### Configuration parameters

| Parameter             | Range      | Description             |
| --------------------- | ---------- | ----------------------- |
| **Temperature**       | 0.0 - 2.0  | Response randomness     |
| **Max Tokens**        | 1+         | Maximum response length |
| **Top P**             | 0.0 - 1.0  | Nucleus sampling        |
| **Presence Penalty**  | -2.0 - 2.0 |                         |
| **Frequency Penalty** | -2.0 - 2.0 |                         |

## Fireworks

Before you use this model, ensure you have a [Fireworks API key](https://fireworks.ai/api-keys).

### Available models

Fireworks provides high-speed inference for popular open-source models and fine-tuned variants, including:

* **Llama:** Meta's Llama models in various sizes.
* **Mixtral:** Mistral's mixture-of-experts models.
* **Qwen:** Alibaba's multilingual models.
* **DeepSeek:** DeepSeek models.
* **Other open models:** Gemma, Phi, and more.

For the current list of available models, refer to [Fireworks' model documentation](https://docs.fireworks.ai/models).

### Configuration parameters

| Parameter       | Range     | Description             |
| --------------- | --------- | ----------------------- |
| **Temperature** | 0.0 - 2.0 | Response randomness     |
| **Max Tokens**  | 1+        | Maximum response length |
| **Top P**       | 0.0 - 1.0 | Nucleus sampling        |

### Tool calling

* **Supported Tool Choices:** `auto`, `required`, `none`.
* **Parallel Execution:** Yes.

## Google Gemini

Before you use this model, ensure you have a [Google AI API key](https://aistudio.google.com/app/apikey).

### Available models

Google offers Gemini models in multiple tiers (Ultra, Pro, Flash) optimized for different use cases.

For the current list of available models, refer to [Google's Gemini documentation](https://ai.google.dev/models/gemini).

### Configuration parameters

| Parameter             | Range     | Description             |
| --------------------- | --------- | ----------------------- |
| **Temperature**       | 0.0 - 2.0 | Response randomness     |
| **Max Output Tokens** | 1+        | Maximum response length |
| **Top P**             | 0.0 - 1.0 | Nucleus sampling        |
| **Top K**             | 1+        | Top-k sampling          |

### Tool calling

* **Supported Tool Choices:** `auto`, `any`, `none`.
* **Parallel Execution:** No.

## Google Vertex AI

Before you use this model, ensure you have [Google Cloud credentials](https://cloud.google.com/vertex-ai/docs/authentication).

### Available models

Google offers Gemini models in multiple tiers (Ultra, Pro, Flash) optimized for different use cases, plus other models available through Vertex AI.

For the current list of available models, refer to the [Vertex AI documentation](https://cloud.google.com/vertex-ai/docs/generative-ai/models).

### Configuration parameters

| Parameter             | Range     | Description             |
| --------------------- | --------- | ----------------------- |
| **Temperature**       | 0.0 - 2.0 | Response randomness     |
| **Max Output Tokens** | 1+        | Maximum response length |
| **Top P**             | 0.0 - 1.0 | Nucleus sampling        |
| **Top K**             | 1+        | Top-k sampling          |

#### Advanced options

* **Region Selection:** Deploy in specific Google Cloud regions.
* **Safety Settings:** Configure content filtering thresholds.

### Tool calling

* **Supported Tool Choices:** `auto`, `any`, `none`.
* **Parallel Execution:** No.

## Groq

Before you use this model, ensure you have a [Groq API key](https://console.groq.com/keys).

### Available models

Groq provides high-speed inference for popular open-source models including Llama, Mixtral, and Gemma variants.

For the current list of available models, refer to [Groq's model documentation](https://console.groq.com/docs/models).

### Configuration parameters

| Parameter       | Range     | Description             |
| --------------- | --------- | ----------------------- |
| **Temperature** | 0.0 - 2.0 | Response randomness     |
| **Max Tokens**  | 1+        | Maximum response length |

### Tool calling

* **Supported Tool Choices:** `auto`, `required`, `none`.
* **Parallel Execution:** Yes.

## Mistral AI

Before you use this model, ensure you have a [Mistral AI API key](https://console.mistral.ai/api-keys/).

### Available models

Mistral offers models in multiple tiers (Large, Medium, Small) optimized for different performance and cost requirements.

For the current list of available models, refer to [Mistral's documentation](https://docs.mistral.ai/platform/endpoints/).

### Configuration parameters

| Parameter       | Range     | Description             |
| --------------- | --------- | ----------------------- |
| **Temperature** | 0.0 - 1.0 | Response randomness     |
| **Max Tokens**  | 1+        | Maximum response length |
| **Top P**       | 0.0 - 1.0 | Nucleus sampling        |

### Tool calling

* **Supported Tool Choices:** `auto`, `any`, `none`.
* **Parallel Execution:** No.

## OpenAI

Before you use this model, ensure you have an [OpenAI API key](https://platform.openai.com/api-keys) or [Azure OpenAI credentials](https://learn.microsoft.com/en-us/azure/ai-services/openai/quickstart).

### Available models

OpenAI offers several model families with different capabilities and price points:

* **GPT series:** General-purpose chat models with various size/capability tiers.
* **o-series:** Reasoning-focused models optimized for complex problem-solving.
* **Legacy models:** Older GPT-3.5 and GPT-4 variants.

For the current list of available models, refer to the [OpenAI documentation](https://platform.openai.com/docs/models).

### Configuration parameters

Standard:

| Parameter             | Range      | Description                                                        |
| --------------------- | ---------- | ------------------------------------------------------------------ |
| **Temperature**       | 0.0 - 2.0  | Controls randomness. Lower = more focused, higher = more creative. |
| **Max Output Tokens** | 1+         | Maximum length of the response                                     |
| **Top P**             | 0.0 - 1.0  | Nucleus sampling threshold. Alternative to temperature.            |
| **Presence Penalty**  | -2.0 - 2.0 | Penalize new topics (positive) or encourage them (negative)        |
| **Frequency Penalty** | -2.0 - 2.0 | Penalize repetition (positive) or allow it (negative)              |
| **Seed**              | Integer    | For reproducible outputs                                           |

Advanced:

**Reasoning Effort**: Available on reasoning-optimized models (o-series and newer GPT models).

Controls reasoning depth before responding. Higher effort = better quality for complex tasks, longer latency.

| Value     | Description                                  |
| --------- | -------------------------------------------- |
| `none`    | Disables reasoning (standard chat behavior)  |
| `minimal` | Minimal reasoning                            |
| `low`     | Light reasoning                              |
| `medium`  | Moderate reasoning (default)                 |
| `high`    | Deep reasoning                               |
| `xhigh`   | Extra deep reasoning (if supported by model) |

<Note>
  When reasoning\_effort is active (not `none`), temperature, top\_p, and penalties are automatically disabled.
</Note>

**Service Tier**: Available on newer models.

Controls request priority and processing allocation.

| Value      | Description                                          |
| ---------- | ---------------------------------------------------- |
| `auto`     | System decides based on load (default)               |
| `default`  | Standard processing queue                            |
| `flex`     | Lower cost, variable latency (if supported by model) |
| `priority` | High-priority queue, lower latency, higher cost      |

**Other parameters:**

* **JSON Mode:** Force valid JSON responses.
* **Responses API:** Improved streaming (default: enabled).
* **Parallel Tool Calls:** Execute multiple tools concurrently.

### Tool calling

* **Supported Tool Choices:** `auto`, `required`, `none`, or specific tool name
* **Parallel Execution:** Yes

## OpenAI Compatible Endpoint

Authentication varies by endpoint (often API key or none).

### Configuration

**Required:**

* **Base URL:** Your endpoint URL (e.g., `https://your-endpoint.com/v1`).
* **Model Name:** Your model identifier.

Works with any framework or service that implements the OpenAI-compatible API format, including:

* Self-hosted open-source inference servers
* Model routing proxies
* Custom model endpoints

### Configuration parameters

All OpenAI-compatible parameters:

| Parameter             | Range      | Description             |
| --------------------- | ---------- | ----------------------- |
| **Temperature**       | 0.0 - 2.0  | Response randomness     |
| **Max Tokens**        | 1+         | Maximum response length |
| **Top P**             | 0.0 - 1.0  | Nucleus sampling        |
| **Frequency Penalty** | -2.0 - 2.0 | Reduce repetition       |
| **Presence Penalty**  | -2.0 - 2.0 | Encourage new topics    |

**Advanced:**

* **JSON Mode:** If endpoint supports it.
* **Streaming:** If endpoint supports it.
* **Function Calling:** If endpoint implements OpenAI format.

### Tool calling

* **Supported Tool Choices:** `auto`, `required`, `none` (if endpoint supports).
* **Parallel Execution:** Yes (if endpoint supports).

### Example endpoints

**Local Ollama:**

```
Base URL: http://localhost:11434/v1
Model: llama3.1
```

**vLLM Server:**

```
Base URL: https://your-server.com/v1
Model: mistral-7b-instruct
```

**LiteLLM Proxy:**

```
Base URL: https://litellm.example.com
Model: gpt-4 (routes to configured backend)
```

## XAI

Before you use this model, ensure you have an [xAI API key](https://console.x.ai/).

### Available models

xAI offers Grok models in multiple sizes for different use cases.

For the current list of available models, refer to [xAI's documentation](https://docs.x.ai/docs).

### Configuration parameters

Standard OpenAI-compatible parameters:

| Parameter             | Range     | Description                |
| --------------------- | --------- | -------------------------- |
| **Temperature**       | 0.0 - 2.0 | Response randomness        |
| **Max Tokens**        | 1+        | Maximum response length    |
| **Top P**             | 0.0 - 1.0 | Nucleus sampling           |
| **Presence Penalty**  | 0 - 2.0   | Hidden on reasoning models |
| **Frequency Penalty** | 0 - 2.0   | Hidden on reasoning models |

### Tool calling

* **Supported Tool Choices:** OpenAI-compatible.
* **Parallel Execution:** Yes (if supported).

## Common Configuration Across All Providers

### Extra Parameters

All providers support a **JSON editor for extra parameters** not exposed in the UI:

```json theme={"theme":{"light":"catppuccin-latte","dark":"catppuccin-mocha"}}
{
  "logprobs": true,
  "top_logprobs": 5,
  "custom_parameter": "value"
}
```

**Use cases:**

* Provider-specific beta features
* Advanced parameters not yet in UI
* Custom metadata for tracking

**Limitation:** Cannot override parameters already in the UI (e.g., can't set temperature here if it's set above)

### Rate Limiting

**Requests Per Second (RPS)** - Available for all providers when running over datasets:

* **Range:** 0 - 500 RPS
* **Purpose:** Respect API rate limits, control costs
* **Default:** Varies by provider

Set this when running experiments or evaluations to avoid hitting rate limits.

## Next steps

<CardGroup cols={2}>
  <Card title="Configure prompt settings" icon="settings" href="/langsmith/managing-model-configurations">
    Learn how to create and manage model configurations in the Playground.
  </Card>

  <Card title="Create a prompt" icon="edit" href="/langsmith/create-a-prompt">
    Get started building prompts with your chosen model provider.
  </Card>
</CardGroup>

***

<div className="source-links">
  <Callout icon="terminal-2">
    [Connect these docs](/use-these-docs) to Claude, VSCode, and more via MCP for real-time answers.
  </Callout>

  <Callout icon="edit">
    [Edit this page on GitHub](https://github.com/langchain-ai/docs/edit/main/src/langsmith/playground-model-providers.mdx) or [file an issue](https://github.com/langchain-ai/docs/issues/new/choose).
  </Callout>
</div>
