Skip to main content

Overview

The Otoroshi LLM Extension provides a complete AI agent framework that can be used within Otoroshi workflows. AI agents are autonomous entities that use LLMs to reason, take actions through tool calls, and collaborate with other agents through handoffs.

What is an AI Agent?

An AI agent is a system that uses a Large Language Model (LLM) as its reasoning engine. Unlike a simple LLM call that produces a single response, an agent operates in a loop:

  1. The agent receives an input (user message)
  2. The LLM processes the input along with system instructions
  3. If the LLM decides it needs more information, it calls tools (functions, MCP connectors, etc.)
  4. The tool results are fed back to the LLM
  5. Steps 3-4 repeat until the LLM produces a final answer or the maximum number of turns is reached

This loop allows agents to perform complex multi-step tasks autonomously.

Key features

  • Tool calling: Agents can use tool functions, MCP connectors, and inline tools defined as workflow nodes
  • Built-in tools: Agents come with a library of ready-to-use tools (filesystem, shell, HTTP, document conversion, tasks, plan, memory, persistent KV, agent delegation) that can be enabled per agent without any external configuration
  • Agent handoffs: An agent can transfer the conversation to another specialized agent
  • Agent spawning: Agents can spawn sub-agents on the fly with custom instructions provided by the LLM at runtime
  • Persistent memory: Agents can remember previous conversations across sessions
  • Working memory (scratchpad): Agents can track tasks, plans, and key-value notes during execution via built-in tools
  • Persistent KV memory: Agents can store and retrieve key-value data that persists across sessions, backed by Redis or PostgreSQL
  • Document conversion: Agents can convert documents (PDF, DOCX, HTML, images) to markdown text via Kreuzberg
  • Guardrails: Input and output validation to ensure safe and compliant responses
  • Configurable turns: Set a maximum number of reasoning loops to control execution
  • LLM-based routing: Use an LLM to intelligently route requests to different workflow paths
  • OpenAI-compatible endpoint: Expose any agent as a standard chat/completions API via the Agent Proxy plugin

Agent components

Workflow nodes

The extension provides three workflow nodes for building agentic workflows:

NodeDescription
AI AgentThe main agent node - executes an agent with tools, handoffs, memory, and guardrails
AI Agent RouterUses an LLM to choose which workflow path to follow
MCP ToolsHelper node to attach an MCP connector to an agent

HTTP plugin

PluginDescription
Agent ProxyExposes an agent as an OpenAI-compatible chat/completions endpoint

Architecture

                    +-------------------+
| User Input |
+--------+----------+
|
+--------v----------+
| AI Agent Node |
| |
| - Instructions |
| - Provider |
| - Tools |
| - Guardrails |
| - Memory |
+--------+----------+
|
+--------v----------+
| Agent Runner |
| (Agent Loop) |
+--------+----------+
|
+--------------+--------------+
| | |
+--------v---+ +------v------+ +----v-------+
| Tool Calls | | Handoffs | | Final |
| (functions,| | (transfer | | Response |
| MCP, etc) | | to another | | |
+--------+---+ | agent) | +------------+
| +------+------+
| |
+--------------+
|
+--------v----------+
| Continue loop with |
| tool results |
+--------------------+

Agent workflow with tools