Agno Runtime
Multi-model runtime with support for all major LLM providers
Claude Code Runtime
Code-specialized runtime optimized for development workflows
Custom Runtimes
Extend with your own framework (LangChain, CrewAI, AutoGen)
What Are Runtimes?
A runtime is the execution layer that sits between your agent’s configuration and the underlying AI model. It handles:- Model Integration: Routing requests to different LLM providers (OpenAI, Anthropic, Google, etc.)
- Tool Execution: Managing Skills and MCP servers for agent capabilities
- Conversation Management: Maintaining context and history across multi-turn interactions
- Streaming & Feedback: Providing real-time execution updates
- Performance Optimization: Caching, batching, and efficient token usage
Quick Comparison
| Feature | Agno | Claude Code |
|---|---|---|
| Framework | Agno + LiteLLM | Claude Code SDK |
| Model Support | All providers | Claude only |
| Best For | Multi-model workflows | Code & development |
| Max History | 100 messages | 200 messages |
| Specialization | Provider flexibility | Development-optimized |
Full Feature Comparison
See detailed side-by-side comparison of all capabilities
Get Started
1
Choose Your Runtime
Determine which runtime best fits your use case:
- Multi-model flexibility needed? → Agno Runtime
- Code generation & analysis? → Claude Code Runtime
- Specialized framework (LangChain, CrewAI)? → Custom Runtime
Compare Runtimes
Use our decision matrix to choose the right runtime
2
Configure Your Agent
Select the runtime when creating an agent:Via CLI:Via API:
Agent Configuration
Learn about agent creation and configuration
3
Execute and Monitor
Your agent will execute using the selected runtime. Monitor performance, token usage, and tool execution through the Kubiya dashboard.
Analytics & Monitoring
Track runtime performance and optimize costs
Built-in Runtimes
Agno Runtime
Multi-model flexibility via LiteLLM
- All LLM providers (GPT, Claude, Gemini, Mistral, Cohere)
- Python-based tool integration
- MCP server support
- 100-message conversation history
- Flexible execution engine
Claude Code Runtime
Code-optimized execution
- Claude models only (optimized integration)
- Advanced file operations
- Repository analysis capabilities
- 200-message extended history
- Session resumption for multi-turn
Extend with Custom Runtimes
Need specialized capabilities? Build your own runtime using popular frameworks:LangChain Integration
LangChain Integration
Integrate LangChain’s ecosystem of tools, chains, and agents. Build custom runtimes that leverage LangChain’s composability while using Kubiya’s orchestration layer.
CrewAI Multi-Agent Systems
CrewAI Multi-Agent Systems
Create runtimes that coordinate multiple specialized agents using CrewAI’s role-based architecture. Perfect for complex workflows requiring agent collaboration.
AutoGen Conversations
AutoGen Conversations
Build runtimes using Microsoft’s AutoGen framework for advanced multi-agent conversations, code execution, and iterative problem-solving.
Custom Framework
Custom Framework
Implement the BaseRuntime interface to integrate any Python-based AI framework. Full control over execution logic, tool integration, and model interaction.
Custom Runtime Developer Guide
Complete guide with code examples, patterns, and best practices
Explore Documentation
Understanding Runtimes
Learn how runtimes work and fit into the Kubiya architecture
Runtime Comparison
Side-by-side feature comparison and decision framework
Custom Runtimes
Build your own runtime with custom frameworks
Control Plane
Runtime registry and orchestration