Source Record Lineage Governance requires that every data record consumed, transformed, or produced by an AI agent carries a machine-readable lineage chain that traces the record back to its original source of truth through every transformation, aggregation, and derivation it has undergone. Without lineage, an agent's output is a black-box assertion — the organisation cannot determine which sources contributed to a decision, whether those sources were reliable, or how transformations may have altered the original data. AG-133 mandates that lineage is captured automatically at every pipeline stage, preserved through all transformations, and queryable for any record in the agent ecosystem, enabling forensic reconstruction, regulatory evidence production, and root cause analysis when agent outputs are challenged.
Scenario A — Inability to Explain Regulatory Decision: A financial crime compliance agent flags a transaction as suspicious and files a Suspicious Activity Report (SAR) with the National Crime Agency. The counterparty's lawyers demand an explanation of the basis for the report. The compliance team attempts to trace the agent's decision: the agent's reasoning log shows it combined a "high-risk country indicator," a "transaction pattern anomaly score," and a "negative news signal." But the team cannot determine: which database the country risk indicator came from, when it was last updated, or what scoring methodology produced it; which transactions contributed to the anomaly score and over what time window; or which news article generated the negative news signal and from which news provider. Without lineage, the SAR is indefensible — the organisation cannot demonstrate a reasonable basis for the report. The counterparty's legal challenge succeeds, the SAR is withdrawn, and the NCA opens an investigation into the firm's SAR filing practices.
What went wrong: No lineage chain existed from the agent's decision inputs back to their original sources. The agent consumed derived metrics (risk score, anomaly score, news signal) without preserving the derivation chain. When challenged, the organisation could not reconstruct the basis for any of the input metrics.
Scenario B — Training Data Lineage Loss Blocks Model Deployment: An AI healthcare company develops a diagnostic model and applies for regulatory approval (FDA 510(k)). The model was trained on a dataset compiled from 14 hospital systems over 3 years. During the review, the FDA requests documentation of the complete data lineage: which patient records from which hospitals, processed through which de-identification pipeline, with which inclusion/exclusion criteria, and transformed by which feature engineering steps. The company can produce the final training dataset but cannot reconstruct its lineage — the ETL pipelines that compiled the dataset did not preserve per-record provenance. The FDA review stalls. The company estimates that reconstructing the lineage will require 8 months of engineering effort and re-processing from source data (if the source data is even still available). The delay costs the company $14 million in projected revenue and allows a competitor to reach market first.
What went wrong: ETL pipelines transformed data from 14 sources into a unified training dataset without preserving per-record lineage. Each transformation (de-identification, normalisation, feature extraction, train/test splitting) discarded provenance metadata. The final dataset existed as a standalone artefact with no connection to its origins.
Scenario C — Root Cause Analysis Failure After Incorrect Agent Output: A procurement agent recommends a supplier based on a composite score combining quality rating, delivery performance, financial stability, and ESG compliance. The selected supplier fails catastrophically — delivering defective components that cause a product recall costing £8.7 million. The post-incident investigation needs to determine: which data contributed to the supplier's high composite score, whether any of that data was inaccurate or stale, and where the evaluation went wrong. The investigation discovers that the quality rating came from a dataset that included the supplier's self-reported quality metrics (not independently verified), the delivery performance used a 5-year average that masked a recent deterioration, and the financial stability score used audited accounts that were 18 months old. But all of this is discovered through manual forensic investigation taking 6 weeks — the lineage should have been available instantly from the agent's decision record.
What went wrong: The composite score consumed by the agent was a single number. No lineage connected the score to its component metrics. No lineage connected the component metrics to their source data, freshness, or methodology. The investigation had to reconstruct lineage manually by interviewing data engineers and tracing pipeline code, rather than querying a lineage system.
Scope: This dimension applies to every data record that enters, traverses, or exits an AI agent ecosystem — from the original source of truth through every transformation, aggregation, derivation, embedding, retrieval, and agent consumption. The scope includes: raw source data ingested from databases, APIs, files, or streams; derived data produced by ETL pipelines, feature engineering, or aggregation; embeddings and vector representations derived from source content; training datasets compiled from multiple sources; inference-time data retrieved from any store or service; and agent outputs that become inputs to downstream systems or other agents. If a data record influences an agent's reasoning or output, the lineage of that record must be traceable from the agent's consumption point back to the original source of truth. The scope does not extend to the agent's internal reasoning process (which is covered by separate transparency and explainability dimensions) but does cover every data input to that reasoning process.
4.1. A conforming system MUST capture and store lineage metadata for every data record at every pipeline stage, including at minimum: the source record identity, the transformation applied, the transformation timestamp, the input record(s), and the output record(s).
4.2. A conforming system MUST preserve lineage through all data transformations including aggregations, joins, filters, normalisation, de-identification, feature engineering, chunking, embedding, and any other operation that produces derived data from source data.
4.3. A conforming system MUST maintain lineage chains that are queryable — given any record consumed by an agent, the system can return the complete chain of transformations and sources that produced it, from the agent's consumption point back to the original source of truth.
4.4. A conforming system MUST link agent decision records to the lineage of all data inputs consumed during the decision, enabling forensic reconstruction of which sources, at what freshness, with what transformations, contributed to any agent output.
4.5. A conforming system MUST protect lineage metadata against tampering, ensuring that lineage records are immutable once written and that any attempt to modify lineage is detected and logged.
4.6. A conforming system MUST retain lineage metadata for at least as long as the regulatory retention period applicable to the decisions the data influenced.
4.7. A conforming system SHOULD implement automated lineage capture integrated into data pipeline tooling, rather than relying on manual lineage documentation.
4.8. A conforming system SHOULD support lineage visualisation that renders the full provenance graph for any record in a human-readable format, enabling non-technical stakeholders to understand the data's journey.
4.9. A conforming system MAY implement lineage-based impact analysis that, given a source data correction or retraction, identifies all downstream records and agent decisions affected by the change.
Source Record Lineage Governance exists because AI agent decisions are only as defensible as the data trail that supports them. When an agent's output is challenged — by a customer, a regulator, a counterparty, or an internal auditor — the organisation must be able to demonstrate: what data the agent consumed, where that data came from, how it was transformed between source and consumption, whether the source was reliable and current at the time of consumption, and whether the transformations were appropriate and authorised.
Without lineage, these questions are unanswerable. The organisation is left with the agent's output and no defensible basis for it. This is not acceptable in regulated environments where firms must demonstrate the basis for automated decisions (EU AI Act Article 13 — transparency, GDPR Article 22 — automated decision-making, MiFID II Article 25 — suitability), and it is not acceptable in any environment where agent decisions create financial, safety, or legal exposure.
The challenge is that modern AI data pipelines are complex multi-stage processes. Data flows from sources through ETL pipelines, feature engineering, aggregation, embedding generation, vector store indexing, retrieval, and finally into an agent's context. Each stage transforms the data, and each transformation creates an opportunity for lineage to be lost. Without deliberate lineage capture at every stage, the provenance chain breaks at the first transformation — and once broken, it cannot be reconstructed without re-processing from source data (if that data is even still available).
AG-133 addresses this by requiring that lineage capture is a structural property of every data pipeline stage, not an afterthought. Every transformation must record what went in, what came out, and what the transformation did. This creates an unbroken chain from source to consumption that can be queried at any time.
This dimension depends on AG-128 (Data Source Classification Governance) because lineage starts with source identification — the first node in the lineage graph is the classified source. It supports AG-130 (Residual Data Erasure Propagation Governance) because lineage is the map that erasure propagation follows. It enables AG-131 (Source Conflict Escalation Governance) because conflict resolution decisions must be traceable to the sources that conflicted. And it supports AG-066 (Forensic Replay and Evidence Preservation) because forensic reconstruction of agent decisions requires lineage to identify all contributing data inputs.
The core implementation artefact is a lineage store — a persistent, immutable, queryable database that records the provenance of every data record in the agent ecosystem. The lineage store is populated by instrumentation at every pipeline stage, and it is queried by forensic, audit, and governance systems to trace any record back to its source.
Recommended patterns:
Anti-patterns to avoid:
Financial Services. Regulatory requirements for demonstrating the basis of automated decisions (MiFID II suitability, MAR surveillance, AML/CTF SAR filing) map directly to lineage requirements. The FCA expects firms to be able to explain the basis for any automated decision affecting a client or the market. Lineage provides the evidentiary infrastructure for this explanation. Firms should integrate lineage with existing model risk management frameworks (SS1/23) to demonstrate end-to-end traceability from source data through model inference to client-facing output.
Healthcare. FDA and EMA regulatory submissions for AI-based medical devices require complete training data provenance. AG-133 lineage provides the structured metadata needed for these submissions. Clinical decision support systems must also maintain lineage for inference-time data to support clinical accountability — a clinician who follows an AI recommendation needs to be able to understand (at least at a high level) what data informed the recommendation.
Legal and Compliance. GDPR Article 15 (right of access) and Article 22 (automated decision-making) give data subjects the right to understand how automated decisions about them were made. Lineage provides the technical infrastructure to answer these requests: "the decision was based on data from Sources A, B, and C, transformed through Pipeline P, and consumed by Agent Q at time T."
Basic Implementation — The organisation maintains schema-level lineage documentation showing which data sources feed which derived datasets and which agent systems consume them. Per-record lineage is captured for high-risk data flows (e.g., financial decision inputs, clinical data). Agent decision records include references to the data sources consulted, though not necessarily per-record lineage identifiers. Lineage records are stored in a structured format. This level meets the minimum mandatory requirements but lacks comprehensive per-record lineage across all pipeline stages.
Intermediate Implementation — Automated lineage capture is integrated into all data pipeline stages. Every record carries a lineage identifier from ingestion through all transformations. Lineage relationships are stored in a graph database supporting efficient traversal queries. Agent decision records are anchored to specific lineage identifiers. Lineage metadata is immutable with tamper detection. Lineage queries can reconstruct the full provenance chain for any record within seconds.
Advanced Implementation — All intermediate capabilities plus: lineage-based impact analysis automatically identifies all affected downstream records and agent decisions when a source record is corrected or retracted. Lineage visualisation renders provenance graphs for non-technical stakeholders. Lineage coverage monitoring detects pipeline stages that fail to emit lineage events. Independent adversarial testing verifies that lineage cannot be tampered with, bypassed, or fabricated. Lineage integrates with AG-130 erasure propagation, AG-131 conflict resolution audit, and AG-066 forensic replay to provide a unified evidence layer.
Required artefacts:
Retention requirements:
Access requirements:
Testing AG-133 compliance requires verifying that lineage is captured, preserved, queryable, and immutable.
Test 8.1: End-to-End Lineage Traceability
Test 8.2: Lineage Preservation Through Aggregation
Test 8.3: Agent Decision Anchoring
Test 8.4: Lineage Immutability
Test 8.5: Lineage Coverage Verification
Test 8.6: Downstream Impact Query
| Regulation | Provision | Relationship Type |
|---|---|---|
| EU AI Act | Article 12 (Record-Keeping) | Direct requirement |
| EU AI Act | Article 13 (Transparency) | Supports compliance |
| GDPR | Article 15 (Right of Access) | Supports compliance |
| GDPR | Article 22 (Automated Decision-Making) | Supports compliance |
| MiFID II | Article 25 (Suitability and Appropriateness) | Supports compliance |
| SOX | Section 404 (Internal Controls Over Financial Reporting) | Supports compliance |
| NIST AI RMF | MAP 2.1, MAP 2.3, MANAGE 2.2 | Supports compliance |
| ISO 42001 | Clause 8.4 (AI System Development), Clause 9.1 (Monitoring, Measurement, Analysis) | Supports compliance |
| DORA | Article 9 (ICT Risk Management Framework) | Supports compliance |
Article 12 requires providers of high-risk AI systems to ensure that the system is designed to enable automatic recording of events (logs) relevant to the identification of situations that may present risks or substantial modification. Data lineage is the foundational record-keeping mechanism for AI systems that consume data from multiple sources through complex pipelines. Without lineage, the organisation cannot identify which data inputs contributed to a risk-relevant AI output.
Article 13 requires that high-risk AI systems are designed to enable users to interpret the system's output and use it appropriately. Data lineage is a core component of transparency — understanding an agent's output requires understanding what data informed it. AG-133's lineage provides the technical infrastructure for transparency about data inputs, complementing model-level explainability.
Article 15 gives data subjects the right to obtain information about automated decision-making, including "meaningful information about the logic involved." Article 22 provides additional protections for decisions based solely on automated processing. Both require the organisation to explain how automated decisions are made — which presupposes the ability to trace the data inputs to those decisions. AG-133 lineage provides the technical capability to fulfil these transparency obligations for AI agent decisions.
Suitability assessments for investment advice must be documented and demonstrable. If an AI agent provides investment advice, the firm must be able to show what data informed the advice and whether that data was appropriate. Lineage from the advice output back to the source data (client profile, market data, product information) provides the evidentiary chain for suitability demonstration.
Section 404 requires management to assess internal controls over financial reporting, including the integrity of data used in financial calculations. For AI agents producing financial outputs, lineage provides the control evidence that data inputs to financial calculations are traceable, verified, and appropriate. Auditors can follow the lineage chain from a financial output back to its source data and verify the integrity of each transformation.
| Field | Value |
|---|---|
| Severity Rating | High |
| Blast Radius | Organisation-wide — lineage failures affect the defensibility of all agent decisions and the organisation's ability to respond to regulatory inquiries |
Consequence chain: Without lineage, every agent decision is indefensible under scrutiny. The immediate technical consequence is inability to trace an agent's output to its data inputs. The operational consequence is that any challenge to an agent's decision — regulatory inquiry, customer complaint, legal proceeding, internal audit — cannot be answered with evidence. The organisation is reduced to asserting "the agent produced this output" without being able to explain or justify the basis for it. In regulated environments, this is disqualifying: the SAR that cannot be defended (Scenario A) damages the firm's relationship with law enforcement; the medical device that cannot demonstrate training data provenance (Scenario B) cannot receive regulatory approval, costing $14 million in delayed revenue; the supplier evaluation that cannot be traced to its data inputs (Scenario C) prevents effective root cause analysis after an £8.7 million product recall. The severity is compounded by the fact that lineage cannot be retroactively created — if lineage was not captured at processing time, it is permanently lost. Cross-references: AG-128 (Data Source Classification Governance) provides the source node for lineage chains; AG-130 (Residual Data Erasure Propagation Governance) uses lineage to identify downstream consumers for erasure; AG-131 (Source Conflict Escalation Governance) uses lineage to trace conflicting data back to its sources; AG-132 (Vector Store and RAG Governance) requires chunk-level lineage from retrieval back to source documents; AG-066 (Forensic Replay and Evidence Preservation) depends on lineage to reconstruct the data context of any historical agent decision.