Provider-agnostic middleware
The following middleware work with any LLM provider:| Middleware | Description |
|---|---|
| Summarization | Automatically summarize conversation history when approaching token limits. |
| Human-in-the-loop | Pause execution for human approval of tool calls. |
| Model call limit | Limit the number of model calls to prevent excessive costs. |
| Tool call limit | Control tool execution by limiting call counts. |
| Model fallback | Automatically fallback to alternative models when primary fails. |
| PII detection | Detect and handle Personally Identifiable Information (PII). |
| To-do list | Equip agents with task planning and tracking capabilities. |
| LLM tool selector | Use an LLM to select relevant tools before calling main model. |
| Tool retry | Automatically retry failed tool calls with exponential backoff. |
| LLM tool emulator | Emulate tool execution using anLLM for testing purposes. |
| Context editing | Manage conversation context by trimming or clearing tool uses. |
Summarization
Automatically summarize conversation history when approaching token limits, preserving recent messages while compressing older context. Summarization is useful for the following:- Long-running conversations that exceed context windows.
- Multi-turn dialogues with extensive history.
- Applications where preserving full conversation context matters.
SummarizationMiddleware
Configuration options
Configuration options
Model for generating summaries. Can be a model identifier string (e.g.,
'openai:gpt-4o-mini') or a BaseChatModel instance. See init_chat_model for more information.Conditions for triggering summarization. Can be:
- A single condition dict (all properties must be met - AND logic)
- A list of condition dicts (any condition must be met - OR logic)
fraction(float): Fraction of model’s context size (0-1)tokens(int): Absolute token countmessages(int): Message count
How much context to preserve after summarization. Specify exactly one of:
fraction(float): Fraction of model’s context size to keep (0-1)tokens(int): Absolute token count to keepmessages(int): Number of recent messages to keep
Custom token counting function. Defaults to character-based counting.
Custom prompt template for summarization. Uses built-in template if not specified. The template should include
{messages} placeholder where conversation history will be inserted.Maximum number of tokens to include when generating the summary. Messages will be trimmed to fit this limit before summarization.
Prefix to add to the summary message. If not provided, a default prefix is used.
Deprecated: Use
trigger: {"tokens": value} instead. Token threshold for triggering summarization.Deprecated: Use
keep: {"messages": value} instead. Recent messages to preserve.Full example
Full example
The summarization middleware monitors message token counts and automatically summarizes older messages when thresholds are reached.Trigger conditions control when summarization runs:
- Single condition object (all properties must be met - AND logic)
- Array of conditions (any condition must be met - OR logic)
- Each condition can use
fraction(of model’s context size),tokens(absolute count), ormessages(message count)
fraction- Fraction of model’s context size to keeptokens- Absolute token count to keepmessages- Number of recent messages to keep
Human-in-the-loop
Pause agent execution for human approval, editing, or rejection of tool calls before they execute. Human-in-the-loop is useful for the following:- High-stakes operations requiring human approval (e.g. database writes, financial transactions).
- Compliance workflows where human oversight is mandatory.
- Long-running conversations where human feedback guides the agent.
HumanInTheLoopMiddleware
Model call limit
Limit the number of model calls to prevent infinite loops or excessive costs. Model call limit is useful for the following:- Preventing runaway agents from making too many API calls.
- Enforcing cost controls on production deployments.
- Testing agent behavior within specific call budgets.
ModelCallLimitMiddleware
Configuration options
Configuration options
Maximum model calls across all runs in a thread. Defaults to no limit.
Maximum model calls per single invocation. Defaults to no limit.
Behavior when limit is reached. Options:
'end' (graceful termination) or 'error' (raise exception)Full example
Full example
The middleware tracks model calls across two scopes:
- Thread limit - Max calls across all runs in a conversation thread (requires checkpointer)
- Run limit - Max calls per single invocation (resets each turn)
'end'- Graceful termination (default)'error'- Raise/throw exception
Tool call limit
Control agent execution by limiting the number of tool calls, either globally across all tools or for specific tools. Tool call limits are useful for the following:- Preventing excessive calls to expensive external APIs.
- Limiting web searches or database queries.
- Enforcing rate limits on specific tool usage.
- Protecting against runaway agent loops.
ToolCallLimitMiddleware
Configuration options
Configuration options
Name of specific tool to limit. If not provided, limits apply to all tools globally.
Maximum tool calls across all runs in a thread (conversation). Persists across multiple invocations with the same thread ID. Requires a checkpointer to maintain state.
None means no thread limit.Maximum tool calls per single invocation (one user message → response cycle). Resets with each new user message.
None means no run limit.Note: At least one of thread_limit or run_limit must be specified.Behavior when limit is reached:
'continue'(default) - Block exceeded tool calls with error messages, let other tools and the model continue. The model decides when to end based on the error messages.'error'- Raise aToolCallLimitExceededErrorexception, stopping execution immediately'end'- Stop execution immediately with aToolMessageand AI message for the exceeded tool call. Only works when limiting a single tool; raisesNotImplementedErrorif other tools have pending calls.
Full example
Full example
Specify limits with:
- Thread limit - Max calls across all runs in a conversation (requires checkpointer)
- Run limit - Max calls per single invocation (resets each turn)
'continue'(default) - Block exceeded calls with error messages, agent continues'error'- Raise exception immediately'end'- Stop with ToolMessage + AI message (single-tool scenarios only)
Model fallback
Automatically fallback to alternative models when the primary model fails. Model fallback is useful for the following:- Building resilient agents that handle model outages.
- Cost optimization by falling back to cheaper models.
- Provider redundancy across OpenAI, Anthropic, etc.
ModelFallbackMiddleware
Configuration options
Configuration options
Full example
Full example
The middleware tries fallback models in order when the primary model fails.
PII detection
Detect and handle Personally Identifiable Information (PII) in conversations using configurable strategies. PII detection is useful for the following:- Healthcare and financial applications with compliance requirements.
- Customer service agents that need to sanitize logs.
- Any application handling sensitive user data.
PIIMiddleware
Configuration options
Configuration options
Type of PII to detect. Can be a built-in type (
email, credit_card, ip, mac_address, url) or a custom type name.How to handle detected PII. Options:
'block'- Raise exception when detected'redact'- Replace with[REDACTED_TYPE]'mask'- Partially mask (e.g.,****-****-****-1234)'hash'- Replace with deterministic hash
Custom detector function or regex pattern. If not provided, uses built-in detector for the PII type.
Check user messages before model call
Check AI messages after model call
Check tool result messages after execution
Full example
Full example
The middleware supports detecting built-in PII types (
email, credit_card, ip, mac_address, url) or custom types with regex patterns.Detection strategies:'block'- Raise exception when detected'redact'- Replace with[REDACTED_TYPE]'mask'- Partially mask (e.g.,****-****-****-1234)'hash'- Replace with deterministic hash
apply_to_input- Check user messages before model callapply_to_output- Check AI messages after model callapply_to_tool_results- Check tool result messages after execution
To-do list
Equip agents with task planning and tracking capabilities for complex multi-step tasks. To-do lists are useful for the following:- Complex multi-step tasks requiring coordination across multiple tools.
- Long-running operations where progress visibility is important.
This middleware automatically provides agents with a
write_todos tool and system prompts to guide effective task planning.TodoListMiddleware
Configuration options
Configuration options
Full example
Full example
Just as humans are more effective when they write down and track tasks, agents benefit from structured task management to break down complex problems.
LLM tool selector
Use an LLM to intelligently select relevant tools before calling the main model. LLM tool selectors are useful for the following:- Agents with many tools (10+) where most aren’t relevant per query.
- Reducing token usage by filtering irrelevant tools.
- Improving model focus and accuracy.
LLMToolSelectorMiddleware
Configuration options
Configuration options
Model for tool selection. Can be a model identifier string (e.g.,
'openai:gpt-4o-mini') or a BaseChatModel instance. See init_chat_model for more information.Defaults to the agent’s main model.Instructions for the selection model. Uses built-in prompt if not specified.
Maximum number of tools to select. Defaults to no limit.
List of tool names to always include in the selection
Full example
Full example
The middleware uses a (typically cheaper) LLM to analyze the user’s query and select the most relevant subset of tools.Benefits:
- Shorter prompts - Reduce complexity by exposing only relevant tools
- Better accuracy - Models choose correctly from fewer options
- Cost savings - Use cheaper model for selection
Tool retry
Automatically retry failed tool calls with configurable exponential backoff. Tool retry is useful for the following:- Handling transient failures in external API calls.
- Improving reliability of network-dependent tools.
- Building resilient agents that gracefully handle temporary errors.
ToolRetryMiddleware
Configuration options
Configuration options
Maximum number of retry attempts after the initial call (3 total attempts with default)
Optional list of tools or tool names to apply retry logic to. If
None, applies to all tools.Either a tuple of exception types to retry on, or a callable that takes an exception and returns
True if it should be retried.Behavior when all retries are exhausted. Options:
'return_message'- Return aToolMessagewith error details (allows LLM to handle failure)'raise'- Re-raise the exception (stops agent execution)- Custom callable - Function that takes the exception and returns a string for the
ToolMessagecontent
Multiplier for exponential backoff. Each retry waits
initial_delay * (backoff_factor ** retry_number) seconds. Set to 0.0 for constant delay.Initial delay in seconds before first retry
Maximum delay in seconds between retries (caps exponential backoff growth)
Whether to add random jitter (
±25%) to delay to avoid thundering herdFull example
Full example
The middleware automatically retries failed tool calls with exponential backoff.Key configuration:
max_retries- Number of retry attempts (default: 2)backoff_factor- Multiplier for exponential backoff (default: 2.0)initial_delay- Starting delay in seconds (default: 1.0)max_delay- Cap on delay growth (default: 60.0)jitter- Add random variation (default: True)
on_failure='return_message'- Return error messageon_failure='raise'- Re-raise exception- Custom callable - Function returning error message
LLM tool emulator
Emulate tool execution using an LLM for testing purposes, replacing actual tool calls with AI-generated responses. LLM tool emulators are useful for the following:- Testing agent behavior without executing real tools.
- Developing agents when external tools are unavailable or expensive.
- Prototyping agent workflows before implementing actual tools.
LLMToolEmulator
Configuration options
Configuration options
List of tool names (str) or BaseTool instances to emulate. If
None (default), ALL tools will be emulated. If empty list, no tools will be emulated.Model to use for generating emulated tool responses. Can be a model identifier string (e.g.,
'openai:gpt-4o-mini') or a BaseChatModel instance. See init_chat_model for more information.Full example
Full example
The middleware uses an LLM to generate plausible responses for tool calls instead of executing the actual tools.
Context editing
Manage conversation context by trimming, summarizing, or clearing tool uses. Context editing is useful for the following:- Long conversations that need periodic context cleanup.
- Removing failed tool attempts from context.
- Custom context management strategies.
ContextEditingMiddleware, ClearToolUsesEdit
Configuration options
Configuration options
List of
ContextEdit strategies to applyToken counting method. Options:
'approximate' or 'model'ClearToolUsesEdit options:Token count that triggers the edit. When the conversation exceeds this token count, older tool outputs will be cleared.
Minimum number of tokens to reclaim when the edit runs. If set to 0, clears as much as needed.
Number of most recent tool results that must be preserved. These will never be cleared.
Whether to clear the originating tool call parameters on the AI message. When
True, tool call arguments are replaced with empty objects.List of tool names to exclude from clearing. These tools will never have their outputs cleared.
Placeholder text inserted for cleared tool outputs. This replaces the original tool message content.
Full example
Full example
The middleware applies context editing strategies when token limits are reached. The most common strategy is
ClearToolUsesEdit, which clears older tool results while preserving recent ones.How it works:- Monitor token count in conversation
- When threshold is reached, clear older tool outputs
- Keep most recent N tool results
- Optionally preserve tool call arguments for context
Provider-specific middleware
These middleware are optimized for specific LLM providers.Anthropic
Middleware specifically designed for Anthropic’s Claude models.| Middleware | Description |
|---|---|
| Prompt caching | Reduce costs by caching repetitive prompt prefixes |
Anthropic prompt caching
Reduce costs by caching repetitive prompt prefixes with Anthropic models. Prompt caching is useful for the following:- Applications with long, repeated system prompts.
- Agents that reuse the same context across invocations.
- Reducing API costs for high-volume deployments.
Learn more about Anthropic prompt caching strategies and limitations.
AnthropicPromptCachingMiddleware
Configuration options
Configuration options
Cache type. Only
'ephemeral' is currently supported.Time to live for cached content. Valid values:
'5m' or '1h'Minimum number of messages before caching starts
Behavior when using non-Anthropic models. Options:
'ignore', 'warn', or 'raise'Full example
Full example
OpenAI
Middleware specifically designed for OpenAI models.Coming soon! Check back for OpenAI-specific middleware optimizations.