See what your AI agent actually did — step by step, decision by decision.
Debug and control autonomous AI systems. Trace every decision, replay any run, visualize the reasoning path.
Traditional LLM tools show you tokens, latency, and prompt logs. But agents are different — they plan, decide, call tools, and adapt. When something goes wrong, the failure isn't in one call. It's in the chain of decisions.
Retrying failed strategies without telling you. Your agent might be stuck in a cycle, burning tokens on the same broken approach.
Returning confident wrong answers. HTTP 200, valid JSON, completely wrong result. Your monitoring sees green.
Producing results you can't reconstruct. When tools chain in ways you didn't anticipate, the output becomes untraceable.
Behaving differently than last week on the same input. Model updates, context changes, and tool responses shift — silently altering behavior.
Referencing data that was never provided. Your agent fills gaps with plausible fiction, making decisions based on information it invented.
Sending entire conversation histories on every call. Costs balloon while context windows overflow — and you can't see where the bloat is.
Your observability tool wasn't built for this.
📡 Opswald Trace
Capture LLM calls, tool invocations, decisions, and observations. Automatic instrumentation via proxy or SDK — no code changes required.
▶️ Opswald Replay
Interactive replay of any agent run. Jump to the exact step where things went wrong. See what the agent knew, what it decided, and why.
🔀 Opswald Graph
Decision flow as a navigable graph. Which observations led to which decisions. Where the agent could have gone a different way.
Add one line to your agent. Every LLM call, tool invocation, and decision is automatically captured.
Or use the proxy — zero code changes. Works with OpenAI, Anthropic, Mistral, local models.
Python SDK · TypeScript SDK · Proxy · LangChain · CrewAI
import opswald
opswald.init(api_key="your-key")
# Every agent call is now traced
# automatically. # Or use the proxy — zero code changes:
OPENAI_BASE_URL=https://proxy.opswald.com/v1 "My agent gave the wrong answer. What happened?"
Open the run, replay it step by step, see exactly where the reasoning went wrong.
"Our agent started failing after Friday's deploy. Why?"
Compare runs before and after. The decision graph shows which path changed.
"We have 20 agents in production. Which ones are breaking?"
See all runs, filter by failures, spot patterns across agents.
"Can we actually trust our agents to run autonomously?"
Not yet. But with full traces and replay, you'll know exactly when you can.
Today: understand what happened.
Tomorrow: control what happens next.
Connect via proxy. See your first trace. No credit card.
Read the Docs →