destruction-layer
Why Observability Tools Fall Short of AI Run Identity
Observability tells you how a system is behaving. Identity tells you what a specific execution was. These are different scopes with different structures.
Observability's Scope: Monitoring Running Systems
Observability is the practice of understanding a system's internal state by examining its external outputs — logs, metrics, and traces. The goal is to answer questions about a running system without modifying it. Is it healthy? Is it degrading? Where is latency accumulating? Which error rates are climbing?
Modern observability platforms do this well. They aggregate signals from across distributed systems. They correlate metrics with traces. They surface anomalies. They enable teams to diagnose problems in production without reproducing them locally.
The unit of analysis in observability is the system over time. Dashboards show trends. Alerts fire on thresholds. Investigations start from a symptom and narrow to a cause. The system is the subject. Its behavior over time is the object of study.
Identity's Scope: Binding Declarations to Specific Executions
Identity does not monitor a system. It defines a single execution. The unit of analysis is not the system over time. It is one run, at one moment, with one composition.
Identity would bind a declaration — what this run is composed of — to a specific execution event. The binding would occur at the point of execution. The record would travel with the output. A third party could verify the record without access to the system that produced it.
This is not what observability does. Observability monitors aggregates. Identity would define instances. Observability is continuous. Identity would be point-in-time. Observability is system-scoped. Identity would be run-scoped.
The structural difference is not subtle. It is categorical. The two activities have different subjects, different temporal characteristics, different verification models, and different consumers.
Why Scope Mismatch Means No Amount of Observability Investment Closes the Gap
More dashboards do not produce run identity. More metrics do not produce run identity. More sophisticated anomaly detection does not produce run identity. Each of these investments improves observability. None changes its scope.
Consider what observability can tell you about an AI run. It can tell you that a run occurred at a particular time. It can tell you which endpoint was called. It can tell you the latency, the token count, the response code. It can tell you whether the model service was experiencing elevated error rates at that time. It can show you the run in the context of system-wide behavior patterns.
It cannot tell you what system prompt was active. It cannot tell you what retrieval context was assembled. It cannot tell you which tool definitions were in scope. It cannot tell you whether the model that responded was the model that was intended. It cannot tell you whether the run's composition matched any declared policy.
These are not gaps in the observability platform's feature set. They are outside the scope of what observability is. Observability monitors the system. Identity would define the run. The system contains the run. The run is not the system. Monitoring the container does not define the contents.
What Observability Would Need to Be to Solve This
To establish run identity, an observability tool would need to stop being an observability tool. It would need to shift from monitoring systems to defining executions. It would need to shift from continuous aggregation to point-in-time capture. It would need to shift from system-scoped analysis to run-scoped declaration. It would need to produce records that are verifiable by parties who have no access to the monitored system.
A tool that did all of this would no longer be an observability tool. It would be something else. The characteristics that make observability valuable — continuous monitoring, system-wide aggregation, trend analysis, anomaly detection — are precisely the characteristics that make it unable to establish identity. Identity requires specificity where observability provides breadth. Identity requires independence where observability provides integration. Identity requires declaration where observability provides observation.
Asking observability to establish identity is asking it to become its own opposite. Not an extension. A contradiction.
The Gap Is Not a Gap of Coverage. It Is a Gap of Category.
Coverage gaps can be closed by adding more instrumentation. Category gaps cannot. The difference between observability and identity is not that observability needs more data points. It is that observability and identity are different kinds of activity directed at different kinds of questions.
Observability asks: how is this system behaving? Identity asks: what was this run? The first is a monitoring question. The second is a definitional question. Monitoring answers do not satisfy definitional needs, no matter how precise the monitoring becomes.
This is why the gap persists despite significant investment in observability tooling for AI systems. The investment improves monitoring. The identity gap is unchanged. Teams with excellent observability still cannot answer the question: given this AI output, what exactly was the run that produced it? They can describe the system's state at the time. They cannot define the run's composition at the moment of execution.
The category that would address this question does not exist within observability. It does not exist anywhere in the current stack. The gap is not waiting to be filled by a better version of an existing tool. It is waiting to be recognized as a distinct problem that requires a distinct approach.
Frequently Asked Questions
+Are AI-specific observability platforms different?
AI-specific observability platforms add AI-relevant metrics — token usage, model latency, prompt lengths, evaluation scores. These are valuable additions to the observability signal. They do not change the structural scope of observability itself. The metrics are still system-level, still continuous, still produced by the monitored system, and still unverifiable from outside. AI-specific does not mean identity-capable.
+What about platforms that capture prompts and responses?
Platforms that capture prompts and responses are performing logging within an observability context. The captured data is produced by the executing system. Its completeness cannot be verified from outside. Whether the captured prompt matches what the model actually received depends on the capture point being correct. This is useful data for the operating team. It is not an independently verifiable identity record.
+Does this mean observability has no role in AI accountability?
Observability has an essential role in AI accountability — monitoring system health, detecting anomalies, surfacing performance degradation, enabling incident investigation. These functions support accountability. They do not constitute identity. Both are needed. Neither replaces the other. The mistake is not in using observability. It is in expecting observability to establish something that is outside its structural scope.