← Blog

OBSERVABILITY

The AI Observability Stack in 2026: What's Changed and What's Still Missing

Osarenren I.February 17, 202614 min read

If you're building AI agents in production, you've probably evaluated at least three observability tools in the last six months. Maybe you're running Langfuse for tracing, Helicone for cost tracking, or Arize for evaluation. Maybe you've stitched together a custom stack with OpenTelemetry and Datadog. The tooling has gotten remarkably good — and remarkably crowded.

But here's what I keep running into, both in my own work and in conversations with other AI engineers: even with all these tools, we still can't answer the most important question about our models. We can tell you what went in. We can tell you what came out. We can tell you how long it took and how much it cost. But when someone asks "why did the model do that?" — we're still guessing.

This post maps the AI observability landscape as it stands in early 2026. I'll cover what's changed, what the current tools do well, and — more importantly — what's still missing. Because the gap isn't in logging or cost tracking anymore. It's somewhere deeper.

The Market Has Exploded

Let's start with the numbers. The AI observability market was valued at roughly $2.1 billion in 2025 and is projected to reach $10.7 billion by 2033, growing at a 22.5% CAGR [1]. That's not a niche — it's a category.

The biggest signal of maturity came in January 2026, when ClickHouse acquired Langfuse as part of a $400 million Series D round [2]. Langfuse had grown to over 20,000 GitHub stars and 26 million SDK installs per month by the end of 2025. The acquisition validated what many of us already knew: LLM observability isn't a feature — it's infrastructure.

Meanwhile, the incumbents have moved in. Datadog added native LLM observability with OpenTelemetry GenAI semantic convention support in December 2025 [3]. Elastic published their GenAI observability roadmap in February 2026 [4]. New Relic, Dynatrace, and Splunk have all shipped AI monitoring capabilities. When the enterprise monitoring giants start building, you know the category is real.

But market size and acquisition activity don't tell you whether the tools actually solve the problems you face in production. Let me walk through what they do — and where they stop.

The Current Landscape: A Practitioner's Map

I've spent time with most of the major platforms over the past year. Here's how I'd categorize them, based on what they actually do well rather than what their marketing pages claim.

ToolPrimary StrengthOpen SourceBest ForStarting Price
LangfuseTracing + prompt managementYesTeams wanting data ownershipFree (self-hosted)
Arize PhoenixML + LLM evaluationYesData science-led teamsFree (open source)
LangSmithLangChain debuggingNoLangChain-native workflowsFree tier available
BraintrustEval + experimentationNoCI/CD-integrated quality gatesFree (1M spans)
HeliconeProxy-based cost trackingPartialQuick setup, cost visibilityFree (10K req/mo)
Datadog LLMUnified infra + LLM monitoringNoExisting Datadog customers$15/host/mo + usage
Weights & BiasesExperiment trackingNoML training workflowsFree tier available
TrueFoundryAI Gateway + FinOpsNoEnterprise cost controlUsage-based

This isn't exhaustive — there are also Fiddler AI (explainability), Arthur AI (governance), WhyLabs (data health), DeepEval (testing), Maxim AI (quality scoring), and more. The point is: there's no shortage of tools. The question is whether they're solving the right problems.

What the Current Stack Does Well

Credit where it's due. The AI observability ecosystem has made enormous progress in three areas that were genuinely painful just 18 months ago.

1. Trace Logging and Visualization

Every serious platform now captures structured traces across multi-step LLM workflows. You can see the full execution path of an agent — which tools it called, what it retrieved from your vector database, how it composed its final response. Langfuse and LangSmith are particularly strong here, with nested trace visualization that makes debugging agent chains significantly easier than reading raw logs.

The emergence of OpenTelemetry GenAI semantic conventions has been a major standardization win [5]. OpenLLMetry, the open-source library extending OTel for AI workloads, means you can instrument once and send telemetry to whatever backend you prefer — Datadog, Elastic, Jaeger, or your own stack [6]. This is the kind of boring infrastructure work that makes an ecosystem mature.

2. Cost Attribution

Token-level cost tracking has gone from "nice to have" to table stakes. Helicone pioneered the proxy-based approach — route your API calls through their proxy, get cost breakdowns by user, feature, and model with zero code changes. TrueFoundry takes it further with FinOps guardrails that can enforce budget caps in real time. Braintrust tags spending by team and feature for trend analysis.

This matters because LLM costs are genuinely unpredictable. An agent that enters a recursive reasoning loop can burn through your monthly budget in hours. A prompt change that increases output length by 20% compounds across millions of requests. Without per-request cost visibility, you're flying blind on the most volatile line item in your infrastructure budget.

3. Automated Evaluation

The shift from manual spot-checking to automated quality evaluation has been transformative. Arize Phoenix and Braintrust both run evaluation pipelines that score every production response against quality criteria — relevance, faithfulness, hallucination rate, toxicity. Braintrust integrates these evaluations directly into CI/CD pipelines, failing builds when quality scores drop below thresholds.

This is genuinely useful. Before automated evaluation, the feedback loop was: user complains → engineer investigates → maybe finds the issue → deploys a fix → waits for more complaints. Now the loop is: automated scorer detects regression → alert fires → engineer investigates with full trace context. That's a real improvement in mean time to detection.

The Five-Layer Framework

Dotan Horovits, writing about the new observability paradigm for AI workloads, proposed a useful framework for thinking about where observability needs to happen [7]. I've adapted it slightly based on my own experience:

LayerWhat It MonitorsCurrent Tool Coverage
ApplicationUser feedback, session analytics, feature usageStrong — most tools cover this
OrchestrationChain performance, guardrails, prompt caching, routingGood — LangSmith, Langfuse excel here
AgenticAgent communication, tool usage, decision treesImproving — newer tools adding agent-specific views
ModelToken usage, inference latency, errors, costStrong — this is where most tools started
Data/RAGRetrieval quality, embedding drift, chunk relevanceModerate — Arize and specialized tools cover this

Horovits made an important observation: AI workloads flip traditional observability tradeoffs. Traditional microservices handle millions of requests per second with millisecond latencies and kilobyte payloads. LLM workloads handle hundreds to thousands of requests per minute with multi-second latencies and payloads that can reach megabytes for multimodal inputs. This means instrumentation overhead is negligible (you're adding microseconds to a 10-second call), but payload storage and analysis become the bottleneck.

The framework is useful. But it's also incomplete. Because there's a layer missing from every framework I've seen — and it's the one that matters most.

The Missing Layer: What's Happening Inside the Model

Here's the uncomfortable truth about every tool in the landscape table above: they all treat the model as a black box.

Think about what that means in practice. Your agent hallucinates a response that costs you a customer. You open your observability dashboard. You can see the full trace — the user's input, the retrieval results, the prompt template, the model's output. You can see that the model was called with temperature 0.7 and max_tokens 2048. You can see that the response took 3.2 seconds and cost $0.004.

What you cannot see is what happened between the input and the output. You cannot see which internal representations activated. You cannot see whether the model's "safety" features were suppressed. You cannot see whether the hallucination was caused by a retrieval failure that the model tried to compensate for, or by a genuine confusion in the model's learned representations. You have the inputs and outputs of a function, but zero visibility into the function's execution.

Current AI observability is like debugging a program by only looking at function arguments and return values — never stepping into the function body.

This isn't a minor gap. It's the difference between monitoring and understanding. And it has concrete consequences for three critical workflows.

Debugging Without Root Causes

When a traditional software system fails, you trace the error to a specific line of code, understand why it failed, and fix it. When an LLM fails, you see the bad output and... try a different prompt. Maybe you adjust the temperature. Maybe you add more context to the system message. You're not debugging — you're doing trial and error.

Research in mechanistic interpretability has shown that model failures often have specific internal causes. Anthropic's circuit tracing work demonstrated that you can trace a model's output back through specific computational pathways — circuits — that causally produce the behavior [8]. When a model hallucinates, specific features activate (or fail to activate) in predictable ways. But no observability tool gives you access to this information.

Security Without Visibility

Every observability platform offers some form of output scanning — toxicity filters, PII detection, content safety checks. These are useful but fundamentally reactive. They catch bad outputs after the model has already generated them.

What they can't do is detect an attack while it's happening inside the model. Research has shown that jailbroken models have distinct internal activation patterns — specific features that activate during adversarial inputs that don't activate during normal operation [9]. Anthropic's Constitutional Classifiers++ architecture uses representation-level probes to detect jailbreak attempts with significantly higher accuracy than output-based classifiers [10]. But this requires access to model internals during inference — something no current observability tool provides.

Compliance Without Explanations

The EU AI Act requires that high-risk AI systems provide meaningful explanations of their decision-making processes [11]. Current observability tools can show you what the model decided, but not why. They can produce an audit trail of inputs and outputs, but not a causal explanation of how the model arrived at its conclusion.

Mechanistic interpretability offers a path to genuine explanations — tracing from an output back through the specific features and circuits that produced it. But this capability doesn't exist in any production observability tool today.

Why the Gap Exists

If model-level understanding is so important, why hasn't anyone built it? The answer is that it requires a fundamentally different technical approach than traditional observability.

Current tools work by intercepting API calls. They sit between your application and the model provider, logging the request and response. This is elegant and non-invasive — you can add Helicone to your stack by changing a single URL. But it means you only see what crosses the API boundary. The model's internal state never leaves the provider's infrastructure.

Model-level observability requires activation extraction — capturing the internal representations at specific layers during inference and analyzing them in real time. This involves:

  • Hook-based instrumentation — registering forward hooks on specific model layers to capture activations as tensors
  • Sparse autoencoder analysis — decomposing high-dimensional activation vectors into interpretable features using pre-trained SAEs
  • Real-time feature monitoring — tracking which features activate for each request and detecting anomalous patterns
  • Circuit tracing — mapping the causal pathways from input tokens through intermediate features to output tokens

This is computationally expensive, technically complex, and requires access to model weights — which means it works differently for open-weight models (where you control inference) versus API-based models (where you don't). It's a genuinely hard engineering problem, which is why the market has focused on the easier wins first.

What the Complete Stack Looks Like

If I were designing the ideal AI observability stack for a production agent system in 2026, it would have six layers, not five:

LayerWhat It ProvidesCurrent State
ApplicationUser feedback, session analytics, business metricsSolved — multiple good options
OrchestrationChain tracing, guardrail monitoring, prompt managementSolved — Langfuse, LangSmith lead
AgenticMulti-agent coordination, tool usage, decision pathsImproving rapidly
Model (External)Token usage, latency, cost, error ratesSolved — commoditized
Data/RAGRetrieval quality, embedding drift, chunk relevanceModerate — specialized tools exist
Model (Internal)Feature activations, circuit behavior, representation healthMissing — the frontier

That sixth layer — model internals — is where the most valuable signals live. It's where you'd detect a hallucination before it reaches the user, catch a jailbreak attempt by its activation signature rather than its output, and trace a model failure to its root cause in specific learned features rather than guessing at prompt fixes.

What to Use Today

The missing layer doesn't mean the existing tools aren't valuable. Here's my honest recommendation for what to deploy right now, based on team size and needs:

Solo developer or small team: Start with Langfuse (self-hosted, free) for tracing and Helicone (free tier) for cost tracking. You get 80% of the value with zero cost. Add OpenTelemetry instrumentation from day one so you can switch backends later.

Growth-stage team (5-20 engineers): Use Braintrust or LangSmith for integrated tracing + evaluation. The CI/CD integration for quality gates is worth the cost. Add Arize Phoenix if you need drift detection or have ML-heavy workflows.

Enterprise: If you're already on Datadog, their LLM observability add-on gives you unified monitoring without another vendor. If you need on-prem deployment or strict data residency, evaluate TrueFoundry or self-hosted Langfuse. For governance and compliance, add Arthur AI or Fiddler AI.

Everyone: Instrument with OpenTelemetry GenAI semantic conventions regardless of which platform you choose. The standard is maturing rapidly, and vendor-neutral instrumentation protects you from lock-in [5].

What to Watch For

The observability landscape will look different in 12 months. Here's what I'm tracking:

OpenTelemetry GenAI going stable. The semantic conventions are currently in development, with a 2026 roadmap that includes full specification support [12]. When they stabilize, expect rapid adoption across all major platforms and a wave of consolidation among smaller tools that can't keep up.

Interpretability-native observability. The research is there — sparse autoencoders can extract interpretable features from model activations, circuit tracing can map causal pathways, representation probes can detect adversarial inputs. The engineering challenge is making this work at production latency and scale. The teams that solve this will own the most valuable layer of the stack.

Agent-specific observability. As AI agents become more autonomous — making multi-step decisions, calling external tools, coordinating with other agents — the observability requirements change. You need to monitor not just individual LLM calls but entire decision trajectories. LangSmith's agent graphs are an early example, but the tooling needs to evolve significantly as agent architectures grow more complex.

Regulatory pressure. The EU AI Act's transparency requirements will force observability from a "nice to have" to a compliance requirement for any team deploying AI in regulated industries. Tools that can provide genuine explanations — not just audit trails — will have a structural advantage.

The Bottom Line

The AI observability stack in 2026 is genuinely impressive for what it covers. Trace logging, cost tracking, latency monitoring, and automated evaluation are solved problems with multiple good solutions. The market has matured from "do we need this?" to "which tool is best for our use case?"

But the most important question — why did the model do that? — remains unanswered. Every tool in the ecosystem treats the model as a black box, logging inputs and outputs without visibility into the computational process that connects them. This isn't a criticism of the tools; it's a reflection of where the technology is. Model-level understanding requires fundamentally different techniques — activation extraction, feature analysis, circuit tracing — that are still transitioning from research to production.

That transition is happening. The research foundations are solid, the engineering challenges are tractable, and the market demand is clear. The teams that add this missing layer to their observability stack won't just have better monitoring — they'll have a fundamentally different relationship with their AI systems. One based on understanding, not inference.

At Prysm AI, we're building that missing layer. If you're interested in seeing inside your models — not just around them — join our waitlist.

References

  1. Market.us, "AI in Observability Market Size," 2025. Projected $10.7B by 2033 at 22.5% CAGR. market.us
  2. SiliconANGLE, "Database maker ClickHouse raises $400M, acquires AI observability startup Langfuse," January 16, 2026. siliconangle.com
  3. Datadog, "LLM Observability natively supports OpenTelemetry GenAI Semantic Conventions," December 2025. datadoghq.com
  4. Elastic, "Observability trends for 2026 (Part 2): GenAI and OpenTelemetry," February 2026. elastic.co
  5. OpenTelemetry, "Semantic conventions for generative AI systems," 2025-2026. opentelemetry.io
  6. OpenTelemetry Blog, "AI Agent Observability — Evolving Standards and Best Practices," March 2025. opentelemetry.io
  7. Dotan Horovits, "Observability for AI Workloads: A New Paradigm for a New Era," Medium, January 2026. medium.com
  8. Anthropic, "On the Biology of a Large Language Model," March 2025. Circuit tracing methodology for understanding model internals. anthropic.com
  9. Anthropic, "Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet," 2024. Demonstrated distinct activation patterns for safety-relevant features. transformer-circuits.pub
  10. Anthropic, "Next-generation safety classifiers using representation re-use," 2025. Constitutional Classifiers++ using internal representations for jailbreak detection. anthropic.com
  11. European Commission, "EU AI Act," 2024. Transparency and explainability requirements for high-risk AI systems. artificialintelligenceact.eu
  12. GitHub, "OpenTelemetry Semantic Conventions 2026 Roadmap," Issue #3330, January 2026. github.com