Skills deepagents-implementation
Implements agents using Deep Agents. Use when building agents with create_deep_agent, configuring backends, defining subagents, adding middleware, or setting up human-in-the-loop workflows.
git clone https://github.com/openclaw/skills
T=$(mktemp -d) && git clone --depth=1 https://github.com/openclaw/skills "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/anderskev/deepagents-implementation" ~/.claude/skills/clawdbot-skills-deepagents-implementation && rm -rf "$T"
skills/anderskev/deepagents-implementation/SKILL.mdDeep Agents Implementation
Core Concepts
Deep Agents provides a batteries-included agent harness built on LangGraph:
: Factory function that creates a configured agentcreate_deep_agent- Middleware: Injected capabilities (filesystem, todos, subagents, summarization)
- Backends: Pluggable file storage (state, filesystem, store, composite)
- Subagents: Isolated task execution via the
tooltask
The agent returned is a compiled LangGraph
StateGraph, compatible with streaming, checkpointing, and LangGraph Studio.
Essential Imports
# Core from deepagents import create_deep_agent # Subagents from deepagents import CompiledSubAgent # Backends from deepagents.backends import ( StateBackend, # Ephemeral (default) FilesystemBackend, # Real disk StoreBackend, # Persistent cross-thread CompositeBackend, # Route paths to backends ) # LangGraph (for checkpointing, store, streaming) from langgraph.checkpoint.memory import InMemorySaver from langgraph.checkpoint.postgres import PostgresSaver from langgraph.store.memory import InMemoryStore # LangChain (for custom models, tools) from langchain.chat_models import init_chat_model from langchain_core.tools import tool
Basic Usage
Minimal Agent
from deepagents import create_deep_agent # Uses Claude Sonnet 4 by default agent = create_deep_agent() result = agent.invoke({"messages": [{"role": "user", "content": "Hello!"}]})
With Custom Tools
from langchain_core.tools import tool from deepagents import create_deep_agent @tool def web_search(query: str) -> str: """Search the web for information.""" return tavily_client.search(query) agent = create_deep_agent( tools=[web_search], system_prompt="You are a research assistant. Search the web to answer questions.", ) result = agent.invoke({"messages": [{"role": "user", "content": "What is LangGraph?"}]})
With Custom Model
from langchain.chat_models import init_chat_model from deepagents import create_deep_agent # OpenAI model = init_chat_model("openai:gpt-4o") # Or Anthropic with custom settings from langchain_anthropic import ChatAnthropic model = ChatAnthropic(model_name="claude-sonnet-4-5-20250929", max_tokens=8192) agent = create_deep_agent(model=model)
With Checkpointing (Persistence)
from langgraph.checkpoint.memory import InMemorySaver from deepagents import create_deep_agent agent = create_deep_agent(checkpointer=InMemorySaver()) # Must provide thread_id with checkpointer config = {"configurable": {"thread_id": "user-123"}} result = agent.invoke({"messages": [...]}, config) # Resume conversation result = agent.invoke({"messages": [{"role": "user", "content": "Follow up"}]}, config)
Streaming
The agent supports all LangGraph stream modes.
Stream Updates
for chunk in agent.stream( {"messages": [{"role": "user", "content": "Write a report"}]}, stream_mode="updates" ): print(chunk) # {"node_name": {"key": "value"}}
Stream Messages (Token-by-Token)
for chunk in agent.stream( {"messages": [{"role": "user", "content": "Explain quantum computing"}]}, stream_mode="messages" ): # Real-time token streaming print(chunk.content, end="", flush=True)
Async Streaming
async for chunk in agent.astream( {"messages": [...]}, stream_mode="updates" ): print(chunk)
Multiple Stream Modes
for mode, chunk in agent.stream( {"messages": [...]}, stream_mode=["updates", "messages"] ): if mode == "messages": print("Token:", chunk.content) else: print("Update:", chunk)
Backend Configuration
StateBackend (Default - Ephemeral)
Files stored in agent state, persist within thread only.
# Implicit - this is the default agent = create_deep_agent() # Explicit from deepagents.backends import StateBackend agent = create_deep_agent(backend=lambda rt: StateBackend(rt))
FilesystemBackend (Real Disk)
Read/write actual files on disk. Enables
execute tool for shell commands.
from deepagents.backends import FilesystemBackend agent = create_deep_agent( backend=FilesystemBackend(root_dir="/path/to/project"), )
StoreBackend (Persistent Cross-Thread)
Uses LangGraph Store for persistence across conversations.
from langgraph.store.memory import InMemoryStore from deepagents.backends import StoreBackend store = InMemoryStore() agent = create_deep_agent( backend=lambda rt: StoreBackend(rt), store=store, # Required for StoreBackend )
CompositeBackend (Hybrid Routing)
Route different paths to different backends.
from langgraph.store.memory import InMemoryStore from deepagents.backends import CompositeBackend, StateBackend, StoreBackend store = InMemoryStore() agent = create_deep_agent( backend=CompositeBackend( default=StateBackend(), # /workspace/* → ephemeral routes={ "/memories/": StoreBackend(store=store), # persistent "/preferences/": StoreBackend(store=store), # persistent }, ), store=store, ) # Files under /memories/ persist across all conversations # Files under /workspace/ are ephemeral per-thread
Subagents
Using the Default General-Purpose Agent
By default, a
general-purpose subagent is available with all main agent tools.
agent = create_deep_agent(tools=[web_search]) # The agent can now delegate via the `task` tool: # task(subagent_type="general-purpose", prompt="Research topic X in depth")
Defining Custom Subagents
from deepagents import create_deep_agent research_agent = { "name": "researcher", "description": "Conducts deep research on complex topics with web search", "system_prompt": """You are an expert researcher. Search thoroughly, cross-reference sources, and synthesize findings.""", "tools": [web_search, document_reader], } code_agent = { "name": "coder", "description": "Writes, reviews, and debugs code", "system_prompt": "You are an expert programmer. Write clean, tested code.", "tools": [code_executor, linter], "model": "openai:gpt-4o", # Optional: different model per subagent } agent = create_deep_agent( subagents=[research_agent, code_agent], system_prompt="Delegate research to the researcher and coding to the coder.", )
Pre-compiled LangGraph Subagents
Use existing LangGraph graphs as subagents.
from deepagents import CompiledSubAgent, create_deep_agent from langgraph.prebuilt import create_react_agent # Existing graph custom_graph = create_react_agent( model="anthropic:claude-sonnet-4-5-20250929", tools=[specialized_tool], prompt="Custom workflow instructions", ) agent = create_deep_agent( subagents=[CompiledSubAgent( name="custom-workflow", description="Runs my specialized analysis workflow", runnable=custom_graph, )] )
Subagent with Custom Middleware
from langchain.agents.middleware import AgentMiddleware class LoggingMiddleware(AgentMiddleware): def transform_response(self, response): print(f"Subagent response: {response}") return response agent_spec = { "name": "logged-agent", "description": "Agent with extra logging", "system_prompt": "You are helpful.", "tools": [], "middleware": [LoggingMiddleware()], # Added after default middleware }
Human-in-the-Loop
Basic Interrupt Configuration
Pause execution before specific tools for human approval.
from deepagents import create_deep_agent agent = create_deep_agent( tools=[send_email, delete_file, web_search], interrupt_on={ "send_email": True, # Simple interrupt "delete_file": True, # Require approval before delete # web_search not listed - runs without approval }, checkpointer=checkpointer, # Required for interrupts )
Interrupt with Options
agent = create_deep_agent( tools=[send_email], interrupt_on={ "send_email": { "allowed_decisions": ["approve", "edit", "reject"] }, }, checkpointer=checkpointer, ) # Invoke - will pause at send_email config = {"configurable": {"thread_id": "user-123"}} result = agent.invoke({"messages": [...]}, config) # Check state state = agent.get_state(config) if state.next: # Has pending interrupt # Resume with approval from langgraph.types import Command agent.invoke(Command(resume={"approved": True}), config) # Or resume with edit agent.invoke(Command(resume={"edited_args": {"to": "new@email.com"}}), config) # Or reject agent.invoke(Command(resume={"rejected": True}), config)
Interrupt on Subagent Tools
# Interrupts apply to subagents too agent = create_deep_agent( subagents=[research_agent], interrupt_on={ "web_search": True, # Interrupt even when subagent calls it }, checkpointer=checkpointer, )
Custom Middleware
Middleware Structure
from langchain.agents.middleware.types import ( AgentMiddleware, ModelRequest, ModelResponse, ) from langchain_core.tools import tool class MyMiddleware(AgentMiddleware): # Tools to inject tools = [] # System prompt content to inject system_prompt = "" def transform_request(self, request: ModelRequest) -> ModelRequest: """Modify request before sending to model.""" return request def transform_response(self, response: ModelResponse) -> ModelResponse: """Modify response after receiving from model.""" return response
Injecting Tools via Middleware
from langchain_core.tools import tool @tool def get_current_time() -> str: """Get the current time.""" from datetime import datetime return datetime.now().isoformat() class TimeMiddleware(AgentMiddleware): tools = [get_current_time] system_prompt = "You have access to get_current_time for time-sensitive tasks." agent = create_deep_agent(middleware=[TimeMiddleware()])
Context Injection Middleware
class UserContextMiddleware(AgentMiddleware): def __init__(self, user_preferences: dict): self.user_preferences = user_preferences @property def system_prompt(self): return f"User preferences: {self.user_preferences}" agent = create_deep_agent( middleware=[UserContextMiddleware({"theme": "dark", "language": "en"})] )
Response Logging Middleware
import logging class LoggingMiddleware(AgentMiddleware): def transform_response(self, response: ModelResponse) -> ModelResponse: logging.info(f"Agent response: {response.messages[-1].content[:100]}...") return response agent = create_deep_agent(middleware=[LoggingMiddleware()])
MCP Tool Integration
Connect MCP (Model Context Protocol) servers to provide additional tools.
from langchain_mcp_adapters.client import MultiServerMCPClient from deepagents import create_deep_agent async def main(): mcp_client = MultiServerMCPClient({ "filesystem": { "command": "npx", "args": ["-y", "@modelcontextprotocol/server-filesystem", "/path"], }, "github": { "command": "npx", "args": ["-y", "@modelcontextprotocol/server-github"], "env": {"GITHUB_TOKEN": os.environ["GITHUB_TOKEN"]}, }, }) mcp_tools = await mcp_client.get_tools() agent = create_deep_agent(tools=mcp_tools) async for chunk in agent.astream( {"messages": [{"role": "user", "content": "List my repos"}]} ): print(chunk)
Additional References
For detailed reference documentation, see:
- Built-in Tools Reference - Complete list of tools available on every agent (filesystem, task management, subagent delegation) with path requirements
- Common Patterns - Production-ready examples including research agents with memory, code assistants with disk access, multi-specialist teams, and production PostgreSQL setup