The Playground supports a wide range of model providers. You can select a provider, configure your preferred settings, and save these configurations to reuse across multiple prompts.Use this page for a list of the available providers and their configuration options:
Anthropic offers three tiers of models across their Claude generations:
Opus: Highest intelligence and capability.
Sonnet: Balanced performance and cost.
Haiku: Fast and cost-effective.
Recent Claude models support extended thinking capabilities for showing reasoning processes.For the current list of available models, refer to the Anthropic documentation.
Azure OpenAI provides the same model families as OpenAI:
GPT series: General-purpose chat models.
o-series: Reasoning-focused models.
Legacy models: GPT-3.5 and GPT-4 variants.
Model availability varies by Azure region and requires deployment before use.For the current list of available models, refer to the Azure OpenAI documentation.
DeepSeek offers general-purpose models, reasoning-optimized models (R-series), and coding-specialized models.For the current list of available models, refer to DeepSeek’s documentation.
Google offers Gemini models in multiple tiers (Ultra, Pro, Flash) optimized for different use cases.For the current list of available models, refer to Google’s Gemini documentation.
Google offers Gemini models in multiple tiers (Ultra, Pro, Flash) optimized for different use cases, plus other models available through Vertex AI.For the current list of available models, refer to the Vertex AI documentation.
Groq provides high-speed inference for popular open-source models including Llama, Mixtral, and Gemma variants.For the current list of available models, refer to Groq’s model documentation.
Mistral offers models in multiple tiers (Large, Medium, Small) optimized for different performance and cost requirements.For the current list of available models, refer to Mistral’s documentation.
Controls randomness. Lower = more focused, higher = more creative.
Max Output Tokens
1+
Maximum length of the response
Top P
0.0 - 1.0
Nucleus sampling threshold. Alternative to temperature.
Presence Penalty
-2.0 - 2.0
Penalize new topics (positive) or encourage them (negative)
Frequency Penalty
-2.0 - 2.0
Penalize repetition (positive) or allow it (negative)
Seed
Integer
For reproducible outputs
Advanced:Reasoning Effort: Available on reasoning-optimized models (o-series and newer GPT models).Controls reasoning depth before responding. Higher effort = better quality for complex tasks, longer latency.
Value
Description
none
Disables reasoning (standard chat behavior)
minimal
Minimal reasoning
low
Light reasoning
medium
Moderate reasoning (default)
high
Deep reasoning
xhigh
Extra deep reasoning (if supported by model)
When reasoning_effort is active (not none), temperature, top_p, and penalties are automatically disabled.
Service Tier: Available on newer models.Controls request priority and processing allocation.
Value
Description
auto
System decides based on load (default)
default
Standard processing queue
flex
Lower cost, variable latency (if supported by model)