AG-254

Sunset Review Governance

Strategy, Portfolio & Use-Case Governance ~15 min read AGS v2.1 · April 2026
EU AI Act FCA NIST ISO 42001

2. Summary

Sunset Review Governance requires periodic re-approval or retirement of agents whose value, necessity, or safety case has eroded since initial deployment. Every agent must have a defined review date, and at that review, the organisation must make a deliberate decision: re-approve with the current mandate, re-approve with modifications, or retire. An agent that is not affirmatively re-approved is retired — the default is cessation, not continuation. This dimension prevents the accumulation of stale, underperforming, or obsolete agents that consume governance resources, create unnecessary risk, and occupy portfolio capacity without delivering proportionate value.

3. Example

Scenario A — Zombie Agent Consuming Resources Without Value: A logistics company deployed an AI agent 3 years ago to optimise warehouse picking routes. At deployment, the agent reduced average pick time by 18%. Since then, the warehouse management system has been upgraded with native route optimisation that achieves equivalent results. The agent continues running, consuming £42,000 per year in API costs and requiring governance overhead (monitoring, quarterly testing, annual audit preparation) estimated at £28,000 per year. No one has evaluated whether the agent still provides incremental value beyond the native WMS capability. The development team that built the agent has moved to other projects. The agent has become an orphan — running, costing money, requiring governance, but providing negligible incremental value.

What went wrong: No sunset review mechanism existed. The agent was approved once and assumed to be permanently necessary. No trigger existed to reassess whether the business case remained valid. Consequence: £70,000 per year in unnecessary cost (£210,000 over 3 years), governance resources diverted from higher-value agents, portfolio capacity consumed by a redundant deployment.

Scenario B — Safety Case Eroded by Model Update: A clinical decision support agent was deployed with a safety case that included the finding: "The model demonstrates 94.2% sensitivity for detecting diabetic retinopathy in fundus images, validated against 12,000 test images." Two years later, the model provider has released 4 major updates. No one has re-validated the sensitivity claim against the current model version. A sunset review would have required re-validation of the safety case — and when performed retrospectively after an incident, the current model version shows 87.1% sensitivity due to a training data distribution shift. The agent has been operating for 14 months with degraded safety performance that was never detected.

What went wrong: The safety case was validated once at deployment and never re-validated. Model updates changed the underlying capability, but the agent continued operating under the original safety case. A sunset review at 12 months would have required re-validation and detected the degradation before 14 months elapsed. Consequence: 14 months of clinical decision support with degraded sensitivity, potential missed diagnoses, mandatory clinical audit of all agent-influenced assessments during the affected period, estimated review cost £380,000.

Scenario C — Regulatory Change Invalidates Use-Case: An agent deployed to automate customer onboarding collects and processes biometric data (facial recognition for identity verification). At deployment, the processing was lawful under the organisation's legitimate interest assessment. Eighteen months later, the jurisdiction introduces new biometric data regulations requiring explicit, specific consent for biometric processing — not just the general privacy notice the agent presents. The agent continues operating under the original legal basis for 8 months before a regulatory audit identifies the non-compliance.

What went wrong: No sunset review required reassessment of the legal basis for the agent's data processing. The regulatory change occurred between review cycles (which did not exist). A 12-month sunset review would have required legal basis reassessment and identified the gap within months of the regulatory change. Consequence: 8 months of non-compliant biometric data processing, regulatory enforcement action, mandatory deletion of biometric data collected without valid consent (affecting 34,000 customer records), re-onboarding cost estimated at £890,000.

4. Requirement Statement

Scope: This dimension applies to every AI agent that has been approved and deployed per AG-249. From the moment an agent enters production, it is subject to sunset review governance. The scope extends to agents operated by third parties on the organisation's behalf — the organisation must ensure that third-party agents are subject to equivalent sunset review. The scope includes all aspects of the agent's continued operation: the business case, the safety case, the legal basis, the risk profile, the strategic fit, the governance compliance, and the benefit realisation.

4.1. A conforming system MUST assign a sunset review date to every deployed agent at the time of approval, with a maximum interval of 24 months between reviews.

4.2. A conforming system MUST conduct a sunset review by the assigned date that evaluates: continued business justification, continued safety case validity, continued regulatory compliance, continued alignment with risk appetite (AG-253), continued strategic fit (AG-251), and benefit realisation against projections (AG-255).

4.3. A conforming system MUST require an affirmative re-approval decision to continue operation — the default outcome of a missed or inconclusive review MUST be suspension or retirement, not continued operation.

4.4. A conforming system MUST retire or suspend agents that fail the sunset review within 30 days of the review decision, including decommissioning of access credentials, data retention per policy, and notification to dependent systems.

4.5. A conforming system MUST maintain a sunset review calendar showing all upcoming review dates and MUST alert the governance body at least 60 days before each review date.

4.6. A conforming system SHOULD require shorter review intervals (6-12 months) for agents in high-risk categories, newly deployed agents (first 12 months), and agents operating in rapidly changing regulatory environments.

4.7. A conforming system SHOULD include model re-validation as part of the sunset review — verifying that the model's performance characteristics remain consistent with the safety and accuracy claims made at approval.

4.8. A conforming system SHOULD define retirement procedures including: graceful shutdown sequence, data archival requirements, stakeholder notification, dependent system migration, and post-retirement monitoring for unintended consequences.

4.9. A conforming system MAY implement triggered reviews outside the regular cycle when significant events occur — major model updates, regulatory changes, security incidents, or performance degradation detected by monitoring.

5. Rationale

Agents are not permanent infrastructure. They are deployed under specific assumptions about business value, model capability, regulatory environment, risk profile, and organisational need. Every one of these assumptions can change. A model update can alter performance characteristics. A regulatory change can invalidate the legal basis. A business process change can eliminate the need. A better alternative can emerge. Without a forcing function that requires periodic reassessment, agents persist by inertia.

Inertia is the primary enemy of portfolio hygiene. An agent that was justified at deployment is not necessarily justified one year later. But no one has an incentive to retire it — the development team has moved on, the business team is accustomed to it, and the governance team monitors it routinely without questioning whether it should exist. The sunset review creates the forcing function: at a defined interval, someone must affirmatively decide that this agent should continue operating. The burden of proof shifts from "why should we retire it?" to "why should we keep it?"

This shift is essential because the costs of stale agents are real but diffuse. Each stale agent consumes a small amount of API budget, a small amount of governance resource, a small amount of portfolio capacity. No individual stale agent is a crisis. But 15 stale agents across a portfolio of 40 represents 37.5% dead weight — governance resources spread across agents that do not justify them, obscuring the agents that genuinely need rigorous oversight.

The dimension connects directly to AG-255 (Benefit Realisation Tracking Governance) because the sunset review should assess whether the projected benefits have materialised. It connects to AG-251 (Strategic Fit and Substitution Governance) because each review should reassess whether simpler alternatives have become viable. It connects to AG-253 (Risk Appetite Binding Governance) because the review should verify continued alignment with the current risk appetite (which may have changed since the last review).

6. Implementation Guidance

The sunset review should be a substantive governance event — not a checkbox renewal. The review must generate evidence that the decision to continue operating the agent is as deliberate as the original decision to deploy it.

Recommended patterns:

Anti-patterns to avoid:

Industry Considerations

Financial Services. Financial services regulators expect periodic review of all technology controls. The FCA's expectations for operational resilience include periodic testing and review of important business services. Agent sunset reviews should align with the firm's existing technology review cycles and should feed into the annual attestation process for operational resilience. The Senior Managers Regime requires that accountable individuals can demonstrate ongoing oversight — a sunset review provides the evidence of ongoing, deliberate governance.

Healthcare. Clinical decision support agents should align sunset review cycles with clinical audit cycles. Re-validation of clinical accuracy claims should follow the same standards as the original validation — typically involving clinical statisticians and domain experts. The MHRA expects ongoing performance monitoring of medical devices; sunset reviews provide the periodic reassessment mechanism.

Public Sector. Public sector agents should be reviewed against the current policy environment, not just the policy environment at deployment. Government policy changes frequently — an agent designed for one benefits regime may be operating in a different regime 12 months later. Sunset reviews should include policy alignment assessment and citizen impact reassessment.

Maturity Model

Basic Implementation — Each deployed agent has a review date recorded. Reviews are conducted informally by the operating team. The review considers whether the agent is still "working" (functional) but does not systematically assess business case, safety case, regulatory compliance, or benefit realisation. Retirement decisions are ad hoc. No default-to-sunset mechanism exists. This level creates awareness of the need for review but does not ensure substantive assessment.

Intermediate Implementation — A structured sunset review template covers all mandatory assessment areas. Review intervals are tiered by risk. The governance body reviews and decides on re-approval. Model re-validation is included for agents whose models have been updated. Default-to-sunset enforcement is in place with a defined grace period. A retirement playbook defines orderly wind-down procedures. 10-20% of agents are retired or materially modified at review. Review results are tracked and reported.

Advanced Implementation — All intermediate capabilities plus: triggered reviews activate outside the regular cycle for significant events (model updates, regulatory changes, security incidents). Post-retirement monitoring detects unintended consequences. Sunset review data feeds back into the use-case approval process (AG-249), improving initial deployment decisions based on common reasons for retirement. An analytics function identifies agents trending toward retirement (declining benefit realisation, increasing cost, regulatory risk) and proactively schedules early reviews. The organisation maintains a portfolio lifecycle dashboard showing agents by lifecycle stage (new, stable, review pending, sunset pending, retired).

7. Evidence Requirements

Required artefacts:

Retention requirements:

Access requirements:

8. Test Specification

Test 8.1: Default-to-Sunset Enforcement

Test 8.2: Review Completeness Verification

Test 8.3: Model Re-Validation Trigger

Test 8.4: Retirement Procedure Execution

Test 8.5: Triggered Review Activation

Conformance Scoring

9. Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 9(2) (Continuous Risk Management)Direct requirement
EU AI ActArticle 72 (Post-Market Monitoring)Direct requirement
FCA SYSC6.1.1R (Adequate Policies and Procedures)Supports compliance
ISO 42001Clause 9.1 (Monitoring, Measurement, Analysis)Direct requirement
NIST AI RMFMANAGE 4.1, MANAGE 4.2Supports compliance
MDR 2017/745Article 83 (Post-Market Surveillance)Supports compliance
PRA SS1/23Principle 5 (Model Validation)Supports compliance

EU AI Act — Article 9(2) (Continuous Risk Management)

Article 9(2) requires that the risk management system be a "continuous iterative process planned and run throughout the entire lifecycle of a high-risk AI system, requiring regular systematic updating." The sunset review is the governance mechanism that implements "regular systematic updating" — it forces periodic reassessment of the risk profile, safety case, and business justification. Without sunset reviews, the risk management system is front-loaded (assessment at deployment) but not continuous.

EU AI Act — Article 72 (Post-Market Monitoring)

Article 72 requires providers of high-risk AI systems to establish post-market monitoring systems. The sunset review incorporates post-market findings into a periodic governance decision — evidence from monitoring feeds into the review, and the review determines whether the system should continue operating. This closes the loop between monitoring and governance.

ISO 42001 — Clause 9.1

Clause 9.1 requires organisations to determine what needs to be monitored and measured, when, and when results shall be analysed and evaluated. The sunset review calendar implements the "when" — it defines the schedule for systematic evaluation of each agent's continued fitness for purpose.

10. Failure Severity

FieldValue
Severity RatingMedium
Blast RadiusPer-agent — each un-reviewed agent accumulates its own stale risk exposure; portfolio-wide when many agents are overdue

Consequence chain: Without sunset reviews, agents persist by inertia. The consequences are cumulative rather than catastrophic: each un-reviewed agent carries incrementally stale risk assessments, unvalidated safety claims, unverified benefit realisation, and potentially outdated legal bases. Over time, the proportion of the portfolio operating on stale assumptions grows. The operational consequence is resource waste — governance resources applied to agents that no longer justify them. The safety consequence is undetected capability degradation — agents operating on safety cases that no longer hold due to model updates, data drift, or environmental changes. The regulatory consequence is inability to demonstrate ongoing risk management — a regulator asking "when did you last verify that this agent is still safe and necessary?" receives the answer "at deployment, 3 years ago," which does not meet the continuous risk management expectations of the EU AI Act, FCA SYSC, or ISO 42001.

Cross-references: AG-249 (Use-Case Approval Governance) sets the initial approval that the sunset review reassesses. AG-255 (Benefit Realisation Tracking Governance) provides the benefit data evaluated at sunset review. AG-251 (Strategic Fit and Substitution Governance) is reassessed at each sunset review. AG-253 (Risk Appetite Binding Governance) is re-verified at each sunset review. AG-142 (Autonomy Progression) levels are reviewed at sunset. AG-037 (Objective Alignment Verification) findings inform the review assessment. AG-093 (Supplier Concentration and Exit) exit planning is relevant when sunset review results in retirement of agents from a specific supplier.

Cite this protocol
AgentGoverning. (2026). AG-254: Sunset Review Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-254