How to Prove AI Agent Decisions for EU AI Act Article 12 Compliance
EU AI Act Article 12 requires automatic logging enabling post-hoc reconstruction of AI system operation. Most teams read that as "add logging." It is not. Logging records that something happened. Proof demonstrates what happened, why, and that the record has not been altered since capture. This article explains what Article 12 actually requires, why standard logs and LLM traces do not satisfy it, and how to implement compliant decision audit trails for high-risk AI systems.
What EU AI Act Article 12 Actually Says
Article 12(1): High-risk AI systems shall technically allow for the automatic recording of events (logs) over the lifetime of the system. Article 12(2): Logging capabilities shall ensure a level of traceability adequate to the purpose and commensurate with the risks. Article 12(3): Logs shall include the period of each use, the reference database against which the input data has been checked, input data, and identification of persons involved. The key phrase is post-hoc reconstruction — not recording that a decision occurred, but recording enough to re-derive what the system did and why.
Which AI Systems Are In Scope (Annex III)
EU AI Act Annex III defines eight high-risk categories: (1) biometric identification, (2) critical infrastructure management, (3) education and vocational training, (4) employment and workers management, (5) access to essential private and public services — including credit scoring, insurance pricing, and medical triage, (6) law enforcement, (7) migration, asylum, and border control, (8) justice and democratic processes. If your AI agent makes decisions in any of these domains affecting EU residents, Article 12 applies.
Why Logs Do Not Prove Decisions
Standard application logs record events — a request occurred, a function was called, a response was returned. They do not record: the reasoning chain the agent used, the options it weighted, the confidence behind the chosen action, the exact context state at decision time, or whether the record has been modified since capture. Without tamper-evidence, a log is a record that something was written, not proof of what the agent decided. Article 12 requires the latter.
Eight Fields Required for Post-Hoc Reconstruction
A compliant EU AI Act Article 12 record must contain: (1) Decision intent — the triggering event and objective. (2) Context snapshot — the exact input state, including retrieved data, at decision time. (3) Reasoning chain — how the agent evaluated the situation. (4) Options considered — what alternatives were weighted. (5) Chosen action and confidence score. (6) Outcome — what the execution produced. (7) Provenance — model version, prompt version, agent ID, timestamp. (8) Cryptographic signature — SHA-256 hash + Ed25519 signature proving the record has not been modified.
Implementation with Tenet AI
Tenet captures all eight fields using the TenetClient intent context manager. Initialize once with your API key. Use tenet.intent() to wrap each decision: call intent.snapshot_context() to capture state, intent.decide() to record options and chosen action, and intent.execute() to record the outcome. Every record is automatically SHA-256 hashed and Ed25519 signed at capture time. Records are stored in an append-only ledger with no DELETE path. Retention policy, replay engine, and compliance PDF export are available out of the box.
LangSmith and Datadog Do Not Satisfy Article 12
LangSmith captures LLM call traces for development debugging — it does not apply cryptographic signing, does not capture context snapshots at the required fidelity, and produces developer-readable output rather than compliance-structured records. Datadog captures infrastructure events — span duration, error rate, memory usage — not decision-level reasoning. Neither tool is designed as a compliance artifact. Using them as Article 12 evidence creates regulatory risk: an auditor or regulator who requests post-hoc reconstruction documentation will find the records insufficient.