airunidentity.com

compliance-blindspot

The AI Identity Compliance Blindspot

Compliance frameworks assume systems can demonstrate what they did. AI systems cannot. The gap is not in the framework. It is in the system.

Claims

What compliance documentation claims to provide

Compliance documentation claims to demonstrate that a system operated under defined, documented conditions. It asserts that controls were in place. It records that processes were followed. It attests that the system behaved within the boundaries set by policy and regulation.

For traditional software, this claim is defensible. The version deployed is known. The configuration is stored in version control. The runtime environment is documented. The inputs and outputs are logged. A reviewer can trace from the compliance claim to the system state and confirm that the claim corresponds to reality.

For AI systems, the claim is made with the same language. The supporting evidence is categorically different. The documentation describes intended conditions. Whether the system operated under those conditions during any specific run is unknown.

Verification Gap

The verification gap

Compliance documentation describes the system as it was configured. It does not describe the system as it executed. These are different things.

A system prompt may be documented in a compliance file. Whether that system prompt was active during a specific run is not confirmed by the documentation. It is assumed. The model version may be specified in a deployment manifest. Whether the provider served that exact version at the moment of execution is not verified. It is trusted.

The retrieval pipeline may be described in an architecture document. Whether the index state, the embedding model, and the similarity threshold matched the documented configuration at the time of a specific run is not recorded. It is inferred.

The verification gap is the distance between what documentation claims and what can be confirmed about a specific execution. In traditional software, this gap is narrow. In AI systems, this gap spans the entire execution. Nothing about the run's actual conditions can be confirmed from documentation alone.

AI vs. Traditional

Why this matters specifically for AI vs. traditional software

Traditional software is static between deployments. The binary does not change. The configuration does not drift. The behavior is determined by code, and the code is versioned. Compliance can reference the version and the configuration, and these references are stable.

AI systems are not static between deployments. Model providers update weights without changing version identifiers. System prompts are modified through feature flags and A/B testing. Retrieval indices are rebuilt on schedules independent of application deployment. Agent behavior depends on runtime decisions that are not predetermined by code.

The compliance model for traditional software assumes a stable mapping between documentation and execution. AI systems break this assumption at every layer. The model may not be what the documentation says. The configuration may have drifted. The context is constructed dynamically. The behavior is emergent.

Compliance for traditional software verifies a snapshot. Compliance for AI systems would need to verify each execution. There is no mechanism to do this. The run has no identity to verify against.

Regulatory Pressure

The growing regulatory expectation and the structural inability to meet it

Regulatory frameworks are expanding to cover AI systems. The EU AI Act requires documentation of AI system behavior. FDA guidance for clinical AI demands evidence of consistent operation. Financial regulators require model risk management that includes execution-level audit capability.

These requirements assume a capability that does not exist. They assume that AI systems can demonstrate, at the level of individual executions, that documented conditions were met. No AI system in production can do this. Not because the regulations are unreasonable. Because the infrastructure lacks the primitive that would make compliance demonstrable.

Organizations respond by producing more documentation. More process descriptions. More attestations. More audit reports. The volume of compliance artifacts increases. The verification gap does not decrease. The artifacts describe intentions. They do not reference executions. They cannot reference executions because executions have no identity to reference.

The regulatory expectation and the technical capability are diverging. Regulations demand execution-level accountability. Infrastructure provides system-level documentation. The distance between these two grows with every new regulation and every unidentified run.

Requirements

What verified compliance would require

For compliance to be verified rather than asserted, any valid system must capture the conditions of each execution at the time of execution. Not before. Not after. At the moment the run occurs.

The captured conditions must be immutable. They cannot be retroactively adjusted to match documentation. They must reflect what actually executed, whether or not that matches what was intended.

The captured conditions must be referenceable. A compliance reviewer must be able to point to a specific run and retrieve its execution conditions. Not a summary. Not an aggregation. The specific conditions of that specific run.

The captured conditions must be comparable against documentation. It must be possible to take the compliance documentation — the claimed conditions — and compare them against the actual conditions captured for a run. Where they match, compliance is demonstrated. Where they diverge, the divergence is visible.

None of this is possible without run identity. The identity is the reference point. Without it, there is nothing to capture, nothing to make immutable, nothing to reference, nothing to compare. Compliance remains an assertion. Verification remains impossible.