What is Interlocute?
A deployment surface for addressable AI nodes. Understand what Interlocute is, what problem it solves, and how to think about it.
What Interlocute is
Interlocute is a managed runtime for deploying AI agents as addressable nodes. Each node is a persistent, named endpoint with its own identity, configuration, memory, and usage tracking. You create a node, configure its behavior, and start calling it — Interlocute handles hosting, scaling, observability, and governance.
Think of it as the deployment layer between your application and the LLM. Your app calls a stable endpoint; Interlocute manages the conversation state, tool execution, cost tracking, and policy enforcement that makes AI production-ready.
What problem it solves
Most teams that build with LLMs reach the same wall: the prototype works, but getting to production requires building an entire platform — authentication, state management, cost tracking, rate limiting, tool orchestration, observability. Interlocute is that platform, so you can focus on what your AI does rather than how it runs.
- Reliable deployment — nodes are managed infrastructure with uptime semantics
- Governed execution — policies, quotas, and refusal semantics are built in
- Addressability — every node has a stable identity and a permanent endpoint
- Observability — every request is logged with full token accounting
- Cost control — usage-based pricing with per-node and per-key attribution
What Interlocute is not
- Not just an API wrapper — Interlocute adds identity, state, governance, and observability on top of the underlying LLM
- Not your application logic — your business rules live in your code; Interlocute handles the AI runtime
- Not a framework — there is no SDK to install or library to link; you interact via standard HTTP
Who it's for
Interlocute is built for developers and teams who need to ship AI-powered features into production reliably. If you are building a support chatbot, an internal assistant, a customer-facing AI feature, or an automated intelligence workflow — and you need it to be accountable, observable, and cost-controlled — Interlocute is your deployment surface.
How to think about it
Interlocute sits between protocol semantics and application meaning:
Your application knows what it wants the AI to do. The LLM provider handles raw inference. Interlocute is the governed runtime in between — managing identity, state, policies, tools, memory, and cost attribution.
Next steps
- Core Concepts — learn the nouns: nodes, threads, policies
- Quickstart — make your first API call in 10 minutes