Everyone observes. Only Waxell governs.

Everyone observes. Only Waxell governs.

Know what every agent did. Control what it does next.

Know what every agent did. Control what it does next.

Waxell Observe is an observability and governance SDK for AI agents in production — it captures every LLM call, tool invocation, and agent decision, then enforces runtime policies before the next step executes. Where other tools record what agents did, Waxell changes what agents can do.

Waxell Observe is an observability and governance SDK for AI agents in production — it captures every LLM call, tool invocation, and agent decision, then enforces runtime policies before the next step executes. Where other tools record what agents did, Waxell changes what agents can do.

Auto-instrumentation — two lines

import waxell_observe as waxell

# One line. Every OpenAI / Anthropic / Groq call is now traced.
waxell.init(api_key="wax_sk_...", api_url="https://api.waxell.dev")

Free during beta. 2-line setup.

SOC 2 Ready

Waxell Observe dashboard showing a completed agent run: duration 638.8 seconds, LLM cost $0.6079, 108,041 tokens used, 6 steps completed. Left panel shows execution steps including accounts_selected, search_completed, validation_completed, and gap_fill_completed. Right panel shows an Execution DAG with 5 nodes — a list_building_agent workflow branching into search_contacts_tool, validate_contacts_tool, fill_gaps_tool, and a second validate_contacts_tool.

Auto-instrumentation — two lines

import waxell_observe as waxell

waxell.init(api_key="wax_sk_...", api_url="https://api.waxell.dev")

Built for Operational Experience (OX)

Built for Developer Experience (DX)

Everyone observes. Only Waxell governs.

AI agents make decisions, call tools, and coordinate with other agents. Waxell Observe captures what happens — automatically — so teams can understand, debug, and govern agent behavior in production.

FreE during beta.

pip install in 2 lines.

Auto-instrumentation — two lines

import waxell_observe as waxell

# One line. Every OpenAI / Anthropic / Groq call is now traced.
waxell.init(api_key="wax_sk_...", api_url="https://api.waxell.dev")

Everyone observes. Only Waxell governs.

AI agents make decisions, call tools, and coordinate with other agents. Waxell Observe captures what happens — automatically — so teams can understand, debug, and govern agent behavior in production.

FreE during beta.

pip install in 2 lines.

Auto-instrumentation — two lines

import waxell_observe as waxell

# One line. Every OpenAI / Anthropic / Groq call is now traced.
waxell.init(api_key="wax_sk_...", api_url="https://api.waxell.dev")

Instruments 177 frameworks, LLMs, and vector DBs — zero config.

Instruments 177 frameworks, LLMs, and vector DBs — zero config.

Enforcing what's allowed to happen next.

Other tools show you what happened. Waxell Observe does that too — but it also enforces what's allowed to happen next. A dashboard after the fact is not governance. It's an autopsy.


Waxell introduces dynamic governance: policies that evaluate agent behavior in real time — before execution, between steps, and after completion. When a policy triggers, the agent receives structured feedback: retry, escalate, or halt.


Observability tells you what your agents did. Governance ensures they only do what they should.

Enforcing what's allowed to happen next.

Other tools show you what happened. Waxell Observe does that too — but it also enforces what's allowed to happen next. A dashboard after the fact is not governance. It's an autopsy.


Waxell introduces dynamic governance: policies that evaluate agent behavior in real time — before execution, between steps, and after completion. When a policy triggers, the agent receives structured feedback: retry, escalate, or halt.


Observability tells you what your agents did. Governance ensures they only do what they should.

What Can Waxell Govern?

25+ policy categories. Each one is a class of problem you no longer solve with hope. Configure rules in the dashboard, enforce them during execution.

Audit

Configure logging and compliance. Every decision, every call, every cost — recorded immutably for review.

Kill

Emergency stop controls. Halt any agent, any workflow, immediately. The button you need when autonomy goes wrong.

Rate-Limit

Control how often workflows can run. Prevent runaway loops, enforce cooldowns, throttle expensive operations.


Content

Input/output content scanning and filtering. Block PII, detect prompt injection, redact sensitive data before it leaves your stack.

LLM

Model-specific constraints. Restrict which models an agent can call, set token ceilings per model, enforce provider allowlists.

Safety

Set safety limits and controls. Define boundaries for what agents are allowed to do — and what they must never do.


Control

Flow control and notifications. Define escalation paths, approval gates, and alerting rules for specific agent behaviors.

Operations

Timeouts, retries, and circuit breakers. Define how agents fail — gracefully, with structure, not silently.


Scheduling

Control when workflows can run. Business hours only, maintenance windows, time-zone-aware execution rules.


Cost

Set spending and token limits. Per-agent, per-user, per-session. Budgets that enforce themselves — not spread-sheets you check on Friday.

Quality

Output validation and quality gates. Score outputs automatically, flag low-confidence responses, block inadequate results.

Plus Compliance, Delegation, Identity, Privacy, Reasoning, and More.

Audit

Configure logging and compliance. Every decision, every call, every cost — recorded immutably for review.

Content

Input/output content scanning and filtering. Block PII, detect prompt injection, redact sensitive data before it leaves your stack.

Control

Flow control and notifications. Define escalation paths, approval gates, and alerting rules for specific agent behaviors.

Cost

Set spending and token limits. Per-agent, per-user, per-session. Budgets that enforce themselves — not spread-sheets you check on Friday.

Kill

Emergency stop controls. Halt any agent, any workflow, immediately. The button you need when autonomy goes wrong.

LLM

Model-specific constraints. Restrict which models an agent can call, set token ceilings per model, enforce provider allowlists.

Operations

Timeouts, retries, and circuit breakers. Define how agents fail — gracefully, with structure, not silently.


Quality

Output validation and quality gates. Score outputs automatically, flag low-confidence responses, block inadequate results.

Rate-Limit

Control how often workflows can run. Prevent runaway loops, enforce cooldowns, throttle expensive operations.


Safety

Set safety limits and controls. Define boundaries for what agents are allowed to do — and what they must never do.


Scheduling

Control when workflows can run. Business hours only, maintenance windows, time-zone-aware execution rules.


Plus Compliance, Delegation, Identity, Privacy, Reasoning, and More.

Audit

Configure logging and compliance. Every decision, every call, every cost — recorded immutably for review.


Control

Flow control and notifications. Define escalation paths, approval gates, and alerting rules for specific agent behaviors.

Kill

Emergency stop controls. Halt any agent, any workflow, immediately. The button you need when autonomy goes wrong


Operations

Timeouts, retries, and circuit breakers. Define how agents fail — gracefully, with structure, not silently.


Rate-Limit

Control how often workflows can run. Prevent runaway loops, enforce cooldowns, throttle expensive operations.


Scheduling

Control when workflows can run. Business hours only, maintenance windows, time-zone-aware execution rules.


Content

Input/output content scanning and filtering. Block PII, detect prompt injection, redact sensitive data before it leaves your stack.

Cost

Set spending and token limits. Per-agent, per-user, per-session. Budgets that enforce themselves — not spread-sheets you check on Friday.

LLM

Model-specific constraints. Restrict which models an agent can call, set token ceilings per model, enforce provider allowlists.

Quality

Output validation and quality gates. Score outputs automatically, flag low-confidence responses, block inadequate results.

Safety

Set safety limits and controls. Define boundaries for what agents are allowed to do — and what they must never do.



Plus Compliance, Delegation, Identity, Privacy, Reasoning, and More.


Worried about runaway spend?

Worried about runaway spend?

Cost policies enforce per-agent and per-user budgets in real time. Set a ceiling and forget about it.

Cost policies enforce per-agent and per-user budgets in real time. Set a ceiling and forget about it.

How Do I Add Observability to My Agents?

Two lines of code. Every LLM call, tool invocation, and agent decision — captured automatically from that point on.

As easy as pip install.


Install the SDK. Set your API key. Initialize before your imports. From that point on, every LLM call, tool invocation, and agent decision is captured automatically — with cost, latency, and token counts attached.


No decorators required to start. No wrapper classes. No changes to your agent logic. When you need more structure, add decorators and context managers incrementally. Governance, scoring, and prompt management are there when you're ready.


Works with any Python agent framework. Supports sync, async, Jupyter notebooks, and production servers. If your agent runs Python, you can observe it.

Need agents that stay inside the lines?

Need agents that stay inside the lines?

Safety and content policies scan inputs and outputs in real time. PII detection, prompt injection blocking, and output filtering — enforced, not suggested.

Safety and content policies scan inputs and outputs in real time. PII detection, prompt injection blocking, and output filtering — enforced, not suggested.

From Every Signal to Every Rule

Waxell Observe captures the full anatomy of an agent run — not just the output, but every decision that led to it. LLM calls with tokens, latency, and cost. Routing decisions with the options considered and the choice made. Chain-of-thought reasoning. Retrieval queries and relevance scores. Tool calls with inputs, outputs, and timing. Full execution trees with parent-child span relationships, powered by OpenTelemetry.


None of this stays in a dashboard. It feeds directly into the Waxell governance plane — where policies evaluate agent behavior in real time. Before execution, between steps, and after completion. When a policy triggers, the agent receives structured feedback: retry with adjusted parameters, escalate to a human, or halt.

Observability without governance is an autopsy. Waxell makes it a control system.

Tired of silent failures?

Tired of silent failures?

Operations policies define how agents fail — with timeouts, retries, circuit breakers, and structured fallback. Not silently. Not at 3am.

Operations policies define how agents fail — with timeouts, retries, circuit breakers, and structured fallback. Not silently. Not at 3am.

How Does Waxell Observe Work With Multi-Agent Systems?

One coordinator, three planners, twelve tool calls — and you can see every branch, every handoff, every decision in a single trace.

Production agents don't run alone. A coordinator dispatches to a planner, which spawns researchers, which call tools. Waxell Observe traces the full tree — every child agent, every span, every decision — linked by session and lineage.

Parent-child relationships are detected automatically. Session IDs, user context, and the observe client propagate through nested calls without manual wiring.

Agents stuck in infinite loops?

Agents stuck in infinite loops?

Rate-limit policies prevent runaway execution. Set cooldowns, enforce throttles, and cap invocation frequency — per agent, per workflow, per user.

Rate-limit policies prevent runaway execution. Set cooldowns, enforce throttles, and cap invocation frequency — per agent, per workflow, per user.

What Frameworks Does Waxell Observe Support?

Your framework. Your stack. No rewrites.

Auto-instruments 200+ libraries. Initialize before importing — no code changes required.

LLM PROVIDERS

OpenAI

Anthropic

Azure

+MORE

+MORE

OpenAI

Anthropic

Azure

+MORE

+MORE

OpenAI

Anthropic

+MORE

+MORE

Azure

VECTOR DATABASES

Pinecone

Chroma

+MORE

Pinecone

+MORE

Chroma

Weaviate

+MORE

Pinecone

+MORE

Chroma

+MORE

Weaviate

frameworks

LangChain

LlamaIndex

CrewAI

+MORE

LangChain

+MORE

LlamaIndex

CrewAI

+MORE

LangChain

+MORE

LlamaIndex

+MORE

CrewAI

INFRASTRUCTURE

PostgreSQL

Redis

MongoDB

+MORE

PostgreSQL

+MORE

Redis

MongoDB

+MORE

PostgreSQL

+MORE

Redis

+MORE

MongoDB

Autonomy, observed

Install the SDK, connect to your Waxell instance, and start capturing what your agents actually do.

Autonomy, observed

Install the SDK, connect to your Waxell instance, and start capturing what your agents actually do.

FAQ

What is Waxell Observe?

Waxell Observe is an observability and governance SDK for AI agents running in production. It auto-instruments Python agent frameworks — capturing LLM calls, tool invocations, decisions, and costs — and enforces runtime policies that control what agents are allowed to do next. Two lines of code to initialize; no changes to agent logic required.

How does Waxell Observe differ from LangSmith or Langfuse?

LangSmith and Langfuse capture what agents did — they are observability tools. Waxell Observe does that too, but adds a governance layer: runtime policies that evaluate agent behavior in real time and return structured feedback (retry, escalate, or halt) before the next step executes. The distinction is observability versus governance: recording what happened versus controlling what happens next.

What can Waxell Observe govern?

Waxell Observe supports 25+ policy categories, including Audit, Content, Control, Cost, Kill, LLM, Operations, Quality, Rate-Limit, Safety, and Scheduling. Each category addresses a class of production risk — from runaway spend and prompt injection to silent failures and unauthorized model access. Policies are configured in the dashboard and enforced during execution, not reviewed after the fact.

How does Waxell Observe work with multi-agent systems?

Waxell Observe traces full agent execution trees, not just individual calls. In multi-agent systems, parent-child relationships between agents are detected automatically — every child agent, every spawned workflow, and every tool call is linked by session and lineage. Session IDs and observability context propagate through nested calls without manual wiring.

Does Waxell Observe work with my existing Python agent framework?

Yes. Waxell Observe auto-instruments 200+ libraries — including LangChain, LlamaIndex, CrewAI, OpenAI, Anthropic, and most major vector databases and infrastructure tools. Initialize the SDK before your other imports and instrumentation begins automatically. No decorators required to start, no changes to your agent logic. Works with sync, async, Jupyter notebooks, and production servers. Agentic governance begins from the moment you call waxell.init() — nothing else needs to change.

What does "dynamic governance" mean for AI agents?

Dynamic governance means policies are evaluated in real time during agent execution — not applied as static pre-filters or reviewed after runs complete. It is the core of what separates agentic governance from traditional observability: rather than recording what agents did and alerting afterward, Waxell evaluates each decision point against configured policies before the next step is allowed to proceed. When a policy triggers, the agent receives structured feedback — retry with adjusted parameters, escalate to a human, or halt execution immediately.

Waxell

Waxell provides observability and governance for AI agents in production. Bring your own framework.

© 2026 Waxell. All rights reserved.

Patent Pending.

Waxell

Waxell provides observability and governance for AI agents in production. Bring your own framework.

© 2026 Waxell. All rights reserved.

Patent Pending.

Waxell

Waxell provides observability and governance for AI agents in production. Bring your own framework.

© 2026 Waxell. All rights reserved.

Patent Pending.