Waxell is the governance and observability layer for AI agents — whether you built them yourself or your team just started using Claude Code last Tuesday.
I manage teams that use agents
I build agents
I build High-Stakes Workflows
Free to start. 2-line setup.
SOC 2 Ready
"What are our agents doing?"
With no visibility into tool calls, costs, decisions, or failures. When something breaks, you find out afterward — if at all.
"How do we control them?"
No policy enforcement. No approval workflows. No audit trail. The agent did what the agent wanted to do.
"How do we run them safely at scale?"
No governed execution environment for high-stakes automation. Hope is not an operations strategy.
Waxell answers all three — across agents you didn't build and agents you're running in production.
If you write agent code, Waxell Observe gives you production-grade observability and governance from the first run. If your team runs agents you didn't build — Claude Code, Cursor, custom GPT workflows — Waxell Connect brings them into a governed workspace with zero code changes. If you're building for workflows where wrong is expensive, Waxell Runtime is the execution environment where governance is native to every step.
Connect
Govern agents you don't control.
Your team is already running Claude Code, Cursor, and custom agents. Connect brings them into a governed workspace — no SDK, no code changes, no engineering ticket required.
Agents get an inbox. You get an audit trail. Decisions that need a human get routed to one.
Features:
Agent coordination mesh — register, discover, and govern any third-party or internally built agent.
Inbox and delegation — human-in-the-loop routing for decisions that need a person.
Rug pull detection — alerts when tool descriptions change unexpectedly.
Slack integration — agent activity surfaced directly in your team's existing workflows.
MCP governance layer — policy checks, PII scanning, and audit trails on every MCP tool call.
Observe
Instrument agents you build.
Add waxell.init() before your imports. Everything after that line is observed and governed — automatically. No wrapper classes. No changes to your agent logic.
Features:
Auto-instruments 200+ libraries: LangChain, CrewAI, LlamaIndex, OpenAI, Anthropic, and more.
Full trace trees — LLM reasoning → tool calls → sub-agent delegation, cost attribution at every node.
25+ policy categories enforced during execution: cost budgets, PII detection, kill switches, rate limits, HIPAA / SOC2 / PCI-DSS compliance profiles.
Human-in-the-loop approvals for high-stakes actions.
Works with any Python framework, sync or async.
Runtime
Governed execution for high-risk agent workflows.
For financial automation, healthcare workflows, and infrastructure operations — Runtime is the environment you build inside. Governance isn't layered on top. It's native to every step. Policies gate what the agent is allowed to do before each step runs. There is no gap between execution and enforcement.
Features:
Isolated execution — every run sandboxed with no shared state between workflows or tenants.
Policy enforcement before the first step — 25+ policy categories gate each step before it executes.
Kill switches at every level — stop any agent, workflow, or session immediately.
Durable workflows — checkpoint at every step; resume from the exact point it stopped, not from the beginning.
Built for compliance — HIPAA, SOC 2, and PCI-DSS profiles. Data residency in US or EU.
Companies aren't debating whether to deploy agents. They're already deployed. The governance conversation starts 6–12 months later. Waxell needs to be in the building before that conversation begins.
For a full feature-by-feature breakdown against LangSmith, Datadog LLM Obs, Arize, and MintMCP, see the Waxell comparison page.
Shadow AI is the new shadow IT.
Developers are running Claude Code, Cursor, and custom agents with no organizational oversight.
Regulation is arriving.
EU AI Act enforcement is underway. US executive orders on AI safety are driving enterprise compliance requirements.
The cost surprises are real.
Companies are finding $50K/month LLM bills with no attribution.
MCP adoption is accelerating.
Anthropic, OpenAI, and Google have all adopted MCP.
Waxell is built to grow with your deployment. Each product delivers value on its own — and the path to the next one is direct.
vs. Datadog / New Relic
vs. LangSmith
LangChain-only. If your stack goes beyond it — and most do — LangSmith goes dark. Waxell sees everything.
vs. Building it yourself
Logging is table stakes. Governance isn't. A 25+ category policy engine with compliance profiles, human-in-the-loop approvals, PII scanning, and full cost attribution across a patent-pending runtime platform is months of engineering — before the UI, the audit trail, or multi-tenant isolation.
vs. "We'll add governance later"
Companies that skip governance during initial deployment spend significantly more fixing it later. Connect makes "early" mean "today, with no code changes.



















