Overview
The Otoroshi LLM Extension provides a complete AI agent framework that can be used within Otoroshi workflows. AI agents are autonomous entities that use LLMs to reason, take actions through tool calls, and collaborate with other agents through handoffs.
What is an AI Agent?
An AI agent is a system that uses a Large Language Model (LLM) as its reasoning engine. Unlike a simple LLM call that produces a single response, an agent operates in a loop:
- The agent receives an input (user message)
- The LLM processes the input along with system instructions
- If the LLM decides it needs more information, it calls tools (functions, MCP connectors, etc.)
- The tool results are fed back to the LLM
- Steps 3-4 repeat until the LLM produces a final answer or the maximum number of turns is reached
This loop allows agents to perform complex multi-step tasks autonomously.
Key features
- Tool calling: Agents can use tool functions, MCP connectors, and inline tools defined as workflow nodes
- Built-in tools: Agents come with a library of ready-to-use tools (filesystem, shell, HTTP, document conversion, tasks, plan, memory, persistent KV, agent delegation) that can be enabled per agent without any external configuration
- Agent handoffs: An agent can transfer the conversation to another specialized agent
- Agent spawning: Agents can spawn sub-agents on the fly with custom instructions provided by the LLM at runtime
- Persistent memory: Agents can remember previous conversations across sessions
- Working memory (scratchpad): Agents can track tasks, plans, and key-value notes during execution via built-in tools
- Persistent KV memory: Agents can store and retrieve key-value data that persists across sessions, backed by Redis or PostgreSQL
- Document conversion: Agents can convert documents (PDF, DOCX, HTML, images) to markdown text via Kreuzberg
- Guardrails: Input and output validation to ensure safe and compliant responses
- Configurable turns: Set a maximum number of reasoning loops to control execution
- LLM-based routing: Use an LLM to intelligently route requests to different workflow paths
- OpenAI-compatible endpoint: Expose any agent as a standard
chat/completionsAPI via the Agent Proxy plugin
Agent components
Workflow nodes
The extension provides three workflow nodes for building agentic workflows:
| Node | Description |
|---|---|
| AI Agent | The main agent node - executes an agent with tools, handoffs, memory, and guardrails |
| AI Agent Router | Uses an LLM to choose which workflow path to follow |
| MCP Tools | Helper node to attach an MCP connector to an agent |
HTTP plugin
| Plugin | Description |
|---|---|
| Agent Proxy | Exposes an agent as an OpenAI-compatible chat/completions endpoint |
Architecture
+-------------------+
| User Input |
+--------+----------+
|
+--------v----------+
| AI Agent Node |
| |
| - Instructions |
| - Provider |
| - Tools |
| - Guardrails |
| - Memory |
+--------+----------+
|
+--------v----------+
| Agent Runner |
| (Agent Loop) |
+--------+----------+
|
+--------------+--------------+
| | |
+--------v---+ +------v------+ +----v-------+
| Tool Calls | | Handoffs | | Final |
| (functions,| | (transfer | | Response |
| MCP, etc) | | to another | | |
+--------+---+ | agent) | +------------+
| +------+------+
| |
+--------------+
|
+--------v----------+
| Continue loop with |
| tool results |
+--------------------+
Agent workflow with tools
