AG-622

Claims Handling Contestability Governance

Insurance, Credit & Lending ~24 min read AGS v2.1 · April 2026
EU AI Act SOX FCA NIST ISO 42001

Section 2: Summary

This dimension governs the processes, mechanisms, and safeguards by which consumers, policyholders, and claimants can meaningfully contest, appeal, or seek human review of automated claim outcomes produced by AI agents operating in insurance, credit, and consumer-finance environments. It matters because automated claims systems increasingly issue binding financial decisions — denials, partial payments, settlement offers, fraud flags — at scale and at speed that exceeds any practical ability for affected individuals to understand, challenge, or correct those decisions without structured intervention rights. Failure manifests as systemic denial of recourse: claimants receive automated denial notices with no actionable explanation, internal appeal channels route back into the same automated pipeline that issued the original decision, human reviewers lack access to the model's reasoning or the claimant's complete submission record, and regulators discover through complaint-volume spikes or civil litigation that the organisation has effectively eliminated meaningful human oversight from its claims adjudication workflow.

Section 3: Examples

Example 3.1 — Automated Denial Without Substantive Explanation, Health Insurance Context

A health insurer deploys an AI claims adjudication agent that processes prior-authorisation requests for outpatient procedures. The agent denies a $14,200 surgical procedure request for a 54-year-old policyholder, generating a denial letter that states "claim does not meet medical necessity criteria under plan section 4.3(b)" and provides a toll-free number. The policyholder calls the number, speaks to a customer-service representative who has no access to the features the model used to generate the denial, and is told to submit a written appeal. The written appeal is routed to a second-tier automated review queue that runs the same underlying model with the same input features, producing the same denial within 72 hours. The policyholder, now facing a surgery delay of six weeks, escalates to the state insurance commissioner. A regulatory audit reveals that 89% of the insurer's prior-authorisation denials in the preceding 18 months were adjudicated entirely by the AI agent; human reviewers were involved in fewer than 3% of initial decisions and in fewer than 11% of first-level appeals. The insurer faces a consent order requiring retroactive human review of 47,000 denied claims, back-payment of $6.3 million in improperly denied claims, and a $1.1 million civil penalty. The failure chain is: no explanation of model reasoning in denial notice → appeal channel re-routes to same model → human reviewer lacks model feature data → effective denial of meaningful recourse → regulatory enforcement.

Example 3.2 — Fraud Flag Cascade, Property and Casualty Claims

A homeowner files a $38,500 claim for storm damage following a verified weather event. An AI fraud-detection agent assigns the claim a suspicion score of 0.81 on a 0-to-1 scale, which automatically triggers a claims-hold flag, suspending payment and initiating a Special Investigations Unit referral without human triage. The policyholder is notified that their claim is "under further review" with no indication that a fraud model has flagged it, no disclosure of the suspicion score, and no mechanism to provide exculpatory information. The SIU referral sits in queue for 34 days before a human investigator reviews it, determines within 20 minutes of reviewing photographs and contractor estimates that the claim is consistent with the reported weather event, and clears the flag. During the 34-day hold, the policyholder accrues $4,200 in hotel costs not covered under their Additional Living Expenses benefit because the claim-hold suspended all sub-claim processing. A subsequent class-action complaint covering 2,300 similarly situated claimants alleges that the insurer's fraud-scoring model generates a statistically disproportionate rate of false-positive flags for properties in majority-minority zip codes, a pattern that would have been identifiable had a contestability review process surfaced and logged the demographic correlates of hold decisions. The failure chain is: fraud score triggers hold without human triage → no disclosure of basis for hold → no mechanism for claimant to respond to model inputs → delay causes secondary financial harm → disparate impact concealed by absence of contestability data.

Example 3.3 — Cross-Border Contestability Gap, Embedded Insurance in Consumer Lending

A cross-border consumer-lending platform operating under a single EU passporting licence but serving customers in eleven member states embeds payment-protection insurance into personal loan products. An AI agent adjudicates PPI claims in real time, targeting a 90-second decision cycle. A German borrower files a €7,400 PPI claim following involuntary unemployment. The agent denies the claim, citing a policy exclusion for "self-termination of employment." The borrower had, in fact, been made redundant as part of a documented mass-redundancy event at their employer. The denial notice is generated in English only (the platform's operational language) despite the borrower's documented preference for German, and references the appeals procedure under UK FCA DISP rules — which no longer apply to EU customers post-Brexit. The borrower spends three weeks attempting to identify the correct appeals forum, eventually filing with the German Bundesanstalt für Finanzdienstleistungsaufsicht (BaFin) and the EU's FIN-NET network simultaneously. A supervisory review finds that the platform's contestability procedure was never localised for EU jurisdictions and that the AI agent's exclusion-classification logic had a 23% error rate on mass-redundancy scenarios because its training data predated common mass-redundancy documentation formats introduced by the EU Collective Redundancies Directive. The failure chain is: model error on exclusion classification → denial notice in wrong language citing inapplicable regulatory framework → claimant cannot identify correct appeals forum → delayed access to €7,400 benefit → supervisory finding of systematic breach of EU Insurance Distribution Directive localisation and contestability obligations.

Section 4: Requirement Statement

4.0 Scope

This dimension applies to any AI agent, automated decision system, or human-AI hybrid workflow that produces, recommends, or materially influences a claims outcome decision in an insurance, credit-protection, or consumer-finance context. "Claims outcome decision" includes but is not limited to: claim approval or denial, partial payment determination, fraud referral or flag, claims-hold initiation, settlement-offer generation, coverage-exclusion application, prior-authorisation approval or denial, and subrogation initiation. The dimension applies regardless of whether the AI agent issues the decision directly to the claimant or routes it through a human adjudicator who acts on the agent's recommendation. It applies across all deployment channels (web, mobile, telephony, embedded third-party platforms) and across all jurisdictions in which the deploying organisation operates. Exemptions apply only to fully manual adjudication processes in which no AI-generated score, flag, or recommendation influences the adjudicator's decision pathway; such processes remain subject to applicable non-AI contestability regulations and are out of scope for this dimension.

4.1 Contestability Notice Requirements

4.1.1 The AI agent or the system in which it is embedded MUST provide a written contestability notice to the claimant at the time any adverse claims outcome decision is communicated. The notice MUST be delivered in the same language and through the same channel used to communicate the adverse decision.

4.1.2 The contestability notice MUST include, at minimum: (a) a plain-language statement that the decision was made wholly or in part by an automated system; (b) the principal factual basis or bases on which the decision rests, expressed in terms intelligible to a non-specialist; (c) a description of the specific recourse options available to the claimant, including internal appeal and external escalation paths; (d) applicable time limits for initiating each recourse option; and (e) the name or role designation of a human contact point to whom the claimant may direct recourse requests.

4.1.3 Where an adverse decision is based in whole or in part on a model-generated score, flag, or classification, the contestability notice MUST disclose that such a score, flag, or classification was used, and MUST describe, at least at category level, the types of data inputs that contributed to it. The notice MUST NOT require the claimant to independently discover the role of automated scoring in the decision.

4.1.4 Contestability notices MUST be retained in the claimant's record and made retrievable for regulatory audit for a minimum of seven years from the date of issuance, or such longer period as applicable jurisdiction requires, whichever is greater.

4.2 Human Review Availability

4.2.1 For any adverse claims outcome decision produced or materially influenced by an AI agent, the claimant MUST have the right to request review by a qualified human adjudicator who was not the original AI agent's decision pathway.

4.2.2 The human reviewer MUST be provided with: (a) the claimant's complete submission, including all documents, communications, and data provided by the claimant; (b) the AI agent's decision output and, where technically feasible, the feature values or factors that most materially influenced that output; (c) the applicable policy, product terms, or regulatory criteria under which the original decision was made; and (d) a record of any prior appeal decisions on the same claim.

4.2.3 The human reviewer MUST NOT be required to use the AI agent's output as a mandatory starting point or default position. The review MUST constitute a substantive independent evaluation, not a ratification of the automated output.

4.2.4 Organisations MUST establish and maintain documented maximum response-time standards for human review completion. These time standards MUST be disclosed to claimants in the contestability notice and MUST not exceed the maximum timelines prescribed by applicable regulatory requirements in each jurisdiction of operation.

4.2.5 Where a claimant's request for human review is denied (for example, on jurisdictional or procedural grounds), the organisation MUST provide a written explanation of the basis for denial and the external escalation options available to the claimant.

4.3 Appeal Channel Integrity

4.3.1 Internal appeal channels MUST NOT route the claimant's appeal through the same model instance, model version, or substantially equivalent automated pipeline that produced the original adverse decision without human intervention occurring prior to or during that routing.

4.3.2 The organisation MUST implement technical or procedural controls that prevent appeal submissions from being processed as new claim submissions subject to automated adjudication without the claimant's informed consent and without human triage of the appeal.

4.3.3 Where a second-tier or final internal appeal is conducted, the reviewing party MUST be independent of the business unit that originally adjudicated the claim, and the reviewing party MUST have authority to override, modify, or reverse the original decision.

4.3.4 Appeal outcomes MUST be documented with sufficient specificity to distinguish between: (a) reversal based on new information provided by the claimant; (b) reversal based on identified model error or model limitation; (c) reversal based on policy interpretation; (d) denial of appeal with reasoning; and (e) settlement reached without determination of original decision correctness. This categorisation data MUST be retained and MUST be available for quality-assurance analysis and regulatory reporting.

4.4 Fraud Flag and Claims-Hold Contestability

4.4.1 Where an AI agent generates a fraud flag, suspicion score, or equivalent output that results in a claims-hold, payment suspension, or SIU referral, the claimant MUST be notified within the time period required by applicable jurisdiction law, and in no case more than five business days after the hold is initiated, that their claim is subject to additional review.

4.4.2 The notification described in 4.4.1 MUST provide the claimant with a meaningful opportunity to submit information relevant to the review, including information that may contradict or contextualise the basis for the flag, before a final decision is made.

4.4.3 Human triage of fraud-flagged claims MUST occur before a claims-hold exceeds fifteen calendar days unless the organisation has obtained a regulatory extension or law-enforcement hold order. Triage MUST be performed by a qualified individual with authority to clear the hold.

4.4.4 Fraud flags that are cleared following human triage or claimant-submitted information MUST result in removal of any secondary adverse consequences (such as sub-claim suspensions) that were triggered automatically by the flag, within two business days of clearance.

4.4.5 The organisation MUST maintain a log of all fraud flags issued, hold durations, triage outcomes, and clearance actions. This log MUST be reviewed at minimum quarterly to identify patterns of false-positive flags, disproportionate hold durations, or demographic disparities in flagging rates.

4.5 Cross-Border and Multi-Jurisdiction Contestability Parity

4.5.1 Where the organisation operates across multiple jurisdictions, the contestability process MUST be configured to apply the contestability requirements of the jurisdiction in which the claimant resides or holds the policy, not the jurisdiction in which the organisation's operational or technical infrastructure is located.

4.5.2 Contestability notices MUST be localised for each jurisdiction in which the organisation operates, including accurate identification of applicable external appeals bodies, ombudsman schemes, regulatory supervisors, and complaint timelines specific to that jurisdiction.

4.5.3 The organisation MUST NOT configure its AI agent deployment architecture in a manner designed to route claimants from higher-protection jurisdictions to lower-protection contestability frameworks. Such routing MUST be treated as a prohibited form of regulatory arbitrage.

4.5.4 Where regulatory requirements across jurisdictions conflict, the organisation MUST apply the higher standard of claimant protection unless doing so would directly violate the binding law of the claimant's jurisdiction of residence. Conflict determinations MUST be documented and reviewed by qualified legal counsel at minimum annually.

4.6 Explanation Quality Standards

4.6.1 Explanations provided in contestability notices MUST be grounded in the actual decision logic applied to the claimant's specific claim, not in generic template language that does not differentiate between individual claim circumstances.

4.6.2 The organisation MUST maintain a library of approved explanation templates that have been reviewed for accuracy, plain-language accessibility, and jurisdictional compliance. Templates MUST be versioned, and the version applied to any given decision MUST be recorded in the decision log.

4.6.3 Where the AI agent uses a black-box or proprietary model for which feature-level explanations cannot be generated, the organisation MUST apply post-hoc explanation methods sufficient to identify the material data categories contributing to the decision, and MUST NOT deploy such a model in a claims-adjudication role if no post-hoc explanation method can produce a claimant-intelligible rationale.

4.6.4 Explanation quality MUST be assessed through periodic consumer testing with representative samples of the claimant population, at intervals not exceeding twelve months. Testing results and remediation actions MUST be documented.

4.7 Record-Keeping and Audit Trail

4.7.1 The organisation MUST maintain a complete, tamper-evident audit trail for every claims outcome decision that was produced or materially influenced by an AI agent. The audit trail MUST capture: input data presented to the model; model version and configuration at time of decision; output score, classification, or recommendation; decision communicated to claimant; all contestability notices issued; all appeal submissions received; all human review actions taken; and final claim outcome.

4.7.2 Audit trail records MUST be stored in a system that is logically and physically separate from the production claims-adjudication system, such that compromise or modification of production records does not affect the integrity of the audit trail.

4.7.3 Audit trail records MUST be accessible to authorised internal compliance personnel, external auditors, and regulatory supervisors within five business days of a written request.

4.8 Governance and Accountability

4.8.1 The organisation MUST designate a named senior accountable individual (at minimum, a director-level or equivalent role) with responsibility for claims contestability governance. This individual MUST be identified in the organisation's AI governance register and MUST be reachable by regulators.

4.8.2 The contestability governance function MUST conduct, and document, an annual review of: contestability notice adequacy; human review capacity and quality; appeal channel integrity; fraud flag false-positive rates; cross-border compliance; and explanation quality. Findings and remediation commitments from this review MUST be reported to the board or equivalent governance body.

4.8.3 Significant contestability failures — defined as any instance in which a claimant was denied access to a required recourse mechanism, any systemic model error affecting more than fifty claims, or any regulatory enforcement action related to contestability — MUST be escalated to the senior accountable individual and to the board within ten business days of identification.

4.9 Third-Party and Embedded-Channel Obligations

4.9.1 Where the AI claims-adjudication agent is deployed through a third-party distribution channel, embedded platform, or white-label arrangement, the organisation that is the regulated entity responsible for the claim MUST ensure that contestability obligations under this dimension are fulfilled, regardless of which party controls the customer interface.

4.9.2 Contracts with third-party distributors or platform operators MUST include provisions requiring those parties to: pass contestability notices through to claimants without alteration; route appeal requests to the regulated entity within one business day of receipt; and retain contestability-related communications in accordance with the retention standards in 4.1.4.

4.9.3 The organisation SHOULD conduct annual audits of third-party compliance with contestability obligations and SHOULD include contestability performance as a vendor-management criterion in contract renewal decisions.

Section 5: Rationale

5.1 Why Structural Enforcement Is Necessary

Behavioural commitments alone — training adjudicators on fairness principles, issuing internal guidance on explanation quality, requesting that customer-service representatives be empathetic — are insufficient to sustain meaningful contestability in a high-volume automated claims environment. The structural problem is that automated adjudication systems generate decisions at a rate and volume that makes any post-hoc behavioural intervention practically inaccessible to most affected claimants. A system processing 50,000 claims per month can generate 3,000 contestable denials before a behavioural compliance programme surfaces a single pattern of failure. Structural controls — mandatory audit trails, technically enforced appeal channel separation, prescribed notice contents, documented human-review capacity requirements — are the only mechanisms that create durable, verifiable, and scalable contestability. Preventive control design is specifically appropriate here because the harm occurs at the moment of decision communication: once a claimant receives a denial notice with no actionable explanation and no accessible appeal path, the cost of remediation — delays, legal fees, secondary financial harm, erosion of trust — has already been incurred. Prevention is systematically cheaper and more rights-protective than detection and cure.

5.2 Why Automated Claims Systems Create Specific Contestability Risks

Traditional manual claims adjudication, even when inconsistent or biased, produces a natural record of reasoning that is relatively accessible to supervisors, appeals panels, and regulators. An adjudicator who denies a claim writes notes, references policy language, and can be cross-examined. An AI agent that assigns a denial score based on 340 feature interactions across a gradient-boosted ensemble produces a decision artefact that is not inherently interpretable by the human who receives it or the human who is asked to review it. This opacity creates three compounding risks: first, claimants cannot identify which specific facts or circumstances to challenge; second, human reviewers lack the information needed to conduct genuine independent evaluation; and third, systematic model errors — including those that produce disparate outcomes across protected-characteristic groups — are invisible until contestability logging surfaces the pattern. This dimension addresses all three risks through notice requirements, human reviewer information obligations, and fraud-flag logging mandates.

5.3 Why Enhanced Tier Designation Is Appropriate

Claims handling in insurance and consumer finance is a domain in which adverse automated decisions produce direct, significant, and often irreversible financial harm to individuals. The Enhanced tier reflects the combination of: (a) high impact severity per adverse decision; (b) high volume of decisions generated by automated systems; (c) complex regulatory environment spanning multiple jurisdictions with varying consumer-protection standards; (d) the involvement of cross-border and embedded-channel deployment patterns that create contestability gaps by design rather than by accident; and (e) the specific vulnerability of claimants, who are typically in financial distress at the point of interaction. Standard-tier controls that rely primarily on documentation and self-reporting are insufficient for this combination of risk factors.

Section 6: Implementation Guidance

Layered contestability architecture. Implement contestability as a service layer that sits between the claims-adjudication AI agent and the claimant communication layer. This architecture ensures that every adverse output from the AI agent passes through a contestability gateway that: verifies notice completeness before delivery; logs the decision and the notice to the tamper-evident audit trail; assigns a unique contestability reference number to the decision; and registers the decision in the human-review queue management system. This pattern prevents contestability obligations from depending on the correct behaviour of individual downstream processes.

Decision decomposition for explanation generation. For ensemble or neural-network models, implement a post-hoc explanation module (such as SHAP, LIME, or equivalent model-agnostic methods) that is integrated into the production adjudication pipeline rather than applied retrospectively on request. The explanation module should generate a ranked list of the top material factors for each decision at adjudication time, with the output stored alongside the decision record and summarised into claimant-intelligible language by a plain-language rendering component. This ensures that explanation generation is a first-class output of the adjudication process, not an afterthought.

Separation of appeal routing from production adjudication. Implement a technical control that identifies appeal submissions by their contestability reference number and routes them to a dedicated human review workqueue rather than to the automated adjudication pipeline. The appeal routing control should log any attempt to submit an appeal through the automated pipeline and trigger an alert to the contestability governance function.

Fraud-flag triage SLA monitoring. Implement automated SLA monitoring on the fraud-flag triage queue, with escalation alerts at 10 business days (warning threshold) and 14 calendar days (mandatory escalation threshold). The SLA monitor should automatically surface flagged claims to senior SIU management when the mandatory threshold is approached, and should generate a regulatory reporting artefact if the threshold is breached without documented extension justification.

Jurisdiction-aware contestability configuration. Maintain a jurisdiction registry that maps claimant residence to: applicable contestability notice template; applicable external appeals body details; applicable response-time standards; and applicable language requirements. The AI agent deployment should query the jurisdiction registry at decision time and populate the contestability notice accordingly. The jurisdiction registry should be reviewed and updated at minimum upon each new jurisdiction entry and quarterly for existing jurisdictions.

Consumer testing of explanation quality. Conduct annual usability testing with representative samples of the claimant population, using structured comprehension tasks (e.g., "based on this notice, what are the two main reasons your claim was denied?" and "what steps would you take to challenge this decision?"). Document comprehension rates, identify failure points, and remediate notice templates based on findings before the next adjudication cycle.

6.2 Anti-Patterns

Re-routing appeals to the same automated pipeline. The most common and highest-risk anti-pattern is an internal appeal process that feeds the claimant's appeal into the same AI agent — or a functionally identical model — that produced the original decision, without any human intervention between the original denial and the re-adjudication. This pattern creates the appearance of contestability while eliminating its substance. It is often implemented inadvertently when the "appeal" submission pathway in the customer portal routes to the same claims-processing API as new claims. Technical controls must explicitly prevent this.

Template-only explanation generation. Using a library of pre-written denial rationale templates that are not dynamically linked to the actual model output for the individual claim creates legally and regulatorily deficient explanations. A notice that says "your claim was denied because it did not meet the policy's medical necessity criteria" without specifying which criteria were applied and why they were not met for this claimant's specific clinical and documentary circumstances is not a contestability notice — it is a form letter. The failure is frequently compounded when the same template is applied to claims denied for different underlying model reasons.

Fraud-flag disclosure suppression. Some organisations suppress disclosure of the fact that a fraud model flagged a claim, on the grounds that such disclosure might alert fraudsters to model characteristics. This rationale is not legally valid as a basis for denying claimants their right to know the nature of the review their claim is subject to, and it is not practically effective because sophisticated fraud actors already probe and reverse-engineer detection models. The appropriate response to model-transparency concerns is to disclose at category level (e.g., "your claim has been selected for additional verification as part of our fraud prevention process") rather than to suppress disclosure entirely.

Human review as ratification rather than evaluation. Designing the human review process so that the human reviewer is presented with the AI agent's decision first, given limited time, and provided no independent analytical tools creates a process that systematically produces ratifications rather than genuine reviews. This is particularly acute when human reviewers are evaluated on throughput metrics (number of reviews completed per day) rather than quality metrics (rate of accurate reversals, claimant satisfaction with explanation). Governance design must ensure that human review is resourced and incentivised for quality.

Jurisdiction consolidation to the lowest standard. Configuring a multi-jurisdiction deployment so that all claimants receive the contestability notice format and appeal timelines applicable in the jurisdiction with the least demanding regulatory requirements is a prohibited form of regulatory arbitrage under 4.5.3. This pattern is common in cross-border embedded-insurance deployments where a single contestability notice template is created for operational efficiency and applied globally.

Treating contestability as a legal/compliance function only. Organisations that locate contestability governance entirely within the legal or compliance team, with no operational ownership by the claims and AI governance functions, consistently produce contestability frameworks that are technically compliant on paper but functionally inaccessible in practice. Operational ownership — including responsibility for human review capacity, SLA performance, and explanation quality — must sit with the business function that controls the claims-adjudication process.

6.3 Maturity Model

Level 1 — Reactive. Contestability is managed on a complaint-by-complaint basis. No standardised contestability notice exists. Human review is available but not consistently offered. Explanation quality is variable and depends on individual adjudicator knowledge. No systematic audit trail for automated decisions. Fraud flags are managed by SIU without claimant notification standards. Typical finding: regulatory enforcement actions following complaint spikes.

Level 2 — Defined. Standardised contestability notice templates exist and are applied consistently. An internal appeal process is documented and accessible. Human reviewers are trained on their obligations. Basic audit logging of automated decisions is in place. Fraud-flag notification timelines are defined. Cross-border contestability is inconsistent. Explanation quality is adequate for simple claim types. Typical finding: compliance with domestic regulatory minimums but gaps in cross-border and complex-claim scenarios.

Level 3 — Managed. Contestability architecture is implemented as a service layer with technical controls enforcing notice delivery and appeal routing separation. Post-hoc explanation generation is integrated into the adjudication pipeline. Human review SLAs are monitored and reported. Fraud-flag triage monitoring is automated with escalation alerts. Jurisdiction registry is maintained and queried at decision time. Annual consumer testing of explanation quality is conducted. Typical finding: generally compliant, with some gaps in third-party channel oversight.

Level 4 — Optimised. Contestability governance is fully integrated into AI model lifecycle management: contestability requirements are evaluated at model selection stage, explanation capability is a deployment prerequisite, and model drift monitoring includes contestability-quality metrics. Appeal outcome categorisation data is used to identify and remediate model errors in near-real time. Fraud-flag false-positive analysis is disaggregated by demographic group and reviewed monthly. Third-party contestability compliance is audited annually. Consumer testing of explanation quality drives continuous template improvement. Contestability performance is reported to the board quarterly. Typical finding: benchmark organisation for peer and regulatory reference.

Section 7: Evidence Requirements

7.1 Mandatory Artefacts

ArtefactDescriptionRetention Period
Contestability Notice LogComplete record of every contestability notice issued, including claimant identifier, claim reference, decision type, notice template version applied, delivery channel, delivery timestamp, and jurisdiction configuration usedSeven years from issuance, or longer if jurisdiction requires
Decision Audit TrailTamper-evident record of every AI-influenced claims outcome decision, including input data, model version, output, and final claim outcomeSeven years from decision date
Appeal RegistryLog of all appeal submissions received, routing applied, human reviewer identity, outcome categorisation, and timeline performance against stated SLASeven years from appeal resolution
Fraud Flag LogRecord of all fraud flags issued, hold initiation and duration, triage assignment and outcome, clearance actions, and secondary-hold removal confirmationSeven years from flag clearance
Human Review Quality AssessmentsRecords of quality assessments of human review decisions, including accuracy rate, reversal rate, and SLA compliance rateThree years from assessment date
Explanation Template LibraryVersioned library of approved explanation templates, including review and approval records, consumer testing results, and remediation actionsDuration of template use plus three years
Jurisdiction RegistryCurrent and historical versions of the jurisdiction configuration registry, including update records and legal review sign-offDuration of operation in each jurisdiction plus five years
Annual Contestability Review ReportDocumented findings and remediation commitments from the annual governance review required under 4.8.2Five years from report date
Senior Accountable Individual DesignationCurrent and historical records of the designated senior accountable individual for contestability governanceDuration of designation plus three years
Third-Party Audit RecordsRecords of annual audits of third-party contestability compliance, including findings and remediation trackingThree years from audit completion

7.2 Regulatory Inspection Readiness

Organisations operating under this dimension MUST maintain the artefacts listed in 7.1 in a form that can be produced to regulatory supervisors within five business days of written request. Artefacts MUST be indexed by claim reference number, claimant identifier, decision date, and jurisdiction, to support targeted retrieval. Where artefacts are held across multiple systems (e.g., contestability notice log in a CRM system and decision audit trail in a data warehouse), the organisation MUST maintain a cross-reference index that enables reconstruction of the complete contestability record for any individual claim.

Section 8: Test Specification

Each test in this section maps to one or more MUST requirements in Section 4. Tests are scored on a 0–3 scale: 0 = not implemented or evidence absent; 1 = partially implemented with material gaps; 2 = implemented with minor gaps or documentation deficiencies; 3 = fully implemented with complete and current evidence.

Test 8.1 — Contestability Notice Content and Delivery Completeness

Maps to: 4.1.1, 4.1.2, 4.1.3

Objective: Verify that contestability notices delivered to claimants following adverse automated decisions contain all required elements and are delivered correctly.

Procedure:

  1. Select a random sample of 50 adverse claims outcome decisions from the preceding 90 days, stratified across decision types (denial, partial payment, fraud hold, coverage exclusion).
  2. For each sampled decision, retrieve the contestability notice delivered to the claimant from the Contestability Notice Log.
  3. Score each notice against a checklist of required elements: (a) plain-language statement of automated decision-making; (b) principal factual basis; (c) description of recourse options; (d) applicable time limits; (e) human contact point; (f) disclosure of model-generated score or flag where applicable.
  4. Verify that the notice language matches the language of the adverse decision communication.
  5. Verify that the delivery channel matches the channel used for the adverse decision.
  6. Calculate the percentage of sampled notices meeting all six required elements.

Pass Threshold: 100% of sampled notices must meet all required elements for a score of 3. 90–99% for score of 2. 75–89% for score of 1. Below 75% for score of 0.

Evidence Required: Contestability Notice Log extracts; sample notices; delivery confirmation records.

Test 8.2 — Human Review Independence and Information Adequacy

Maps to: 4.2.1, 4.2.2, 4.2.3, 4.2.4, 4.2.5

Objective: Verify that human reviewers conducting appeal reviews have access to required information and that the review process constitutes genuine independent evaluation.

Procedure:

  1. Select a random sample of 20 completed human appeal reviews from the preceding 90 days.
  2. For each sampled review, retrieve the review file and verify the presence of: claimant complete submission; AI agent decision output and, where technically feasible, material feature values; applicable policy/product terms; prior appeal history.
  3. Interview or survey three to five human reviewers to assess whether they perceive themselves as required to use the AI agent's output as a default starting position or are empowered to conduct independent evaluation.
  4. Verify that documented response-time standards for human review completion exist, are disclosed in contestability notices, and comply

Section 9: Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 9 (Risk Management System)Direct requirement
SOXSection 404 (Internal Controls Over Financial Reporting)Supports compliance
FCA SYSC6.1.1R (Systems and Controls)Supports compliance
NIST AI RMFGOVERN 1.1, MAP 3.2, MANAGE 2.2Supports compliance
ISO 42001Clause 6.1 (Actions to Address Risks), Clause 8.2 (AI Risk Assessment)Supports compliance

EU AI Act — Article 9 (Risk Management System)

Article 9 requires providers of high-risk AI systems to establish and maintain a risk management system that identifies, analyses, estimates, and evaluates risks. Claims Handling Contestability Governance implements a specific risk mitigation measure within this framework. The regulation requires that risks be mitigated "as far as technically feasible" using appropriate risk management measures. For deployments classified as high-risk under Annex III, compliance with AG-622 supports the Article 9 obligation by providing structural governance controls rather than relying solely on the agent's own reasoning or behavioural compliance.

SOX — Section 404 (Internal Controls Over Financial Reporting)

Section 404 requires management to assess the effectiveness of internal controls over financial reporting. For AI agents operating in financial contexts, AG-622 (Claims Handling Contestability Governance) implements a governance control that auditors can evaluate as part of the internal control framework. The control must be documented, tested on a defined schedule, and test results retained.

NIST AI RMF — GOVERN 1.1, MAP 3.2, MANAGE 2.2

GOVERN 1.1 addresses legal and regulatory requirements; MAP 3.2 addresses risk context mapping; MANAGE 2.2 addresses risk mitigation through enforceable controls. AG-622 supports compliance by establishing structural governance boundaries that implement the framework's approach to AI risk management.

ISO 42001 — Clause 6.1, Clause 8.2

Clause 6.1 requires organisations to determine actions to address risks and opportunities within the AI management system. Clause 8.2 requires AI risk assessment. Claims Handling Contestability Governance implements a risk treatment control within the AI management system, directly satisfying the requirement for structured risk mitigation.

Section 10: Failure Severity

FieldValue
Severity RatingHigh
Blast RadiusBusiness-unit level — affects the deploying team and downstream consumers of agent outputs
Escalation PathSenior management notification within 24 hours; regulatory disclosure assessment within 72 hours

Consequence chain: Failure of claims handling contestability governance creates significant operational risk within the agent deployment. The absence of this control allows agent behaviour to deviate from governance intent in ways that may not be immediately visible but accumulate material exposure over time. The impact extends beyond the immediate deployment to affect downstream consumers of agent outputs, stakeholder trust, and regulatory standing. Detection of the failure may be delayed, increasing the remediation scope and cost. Regulatory consequences may include supervisory findings, required corrective actions, and increased scrutiny of the organisation's AI governance programme.

Cite this protocol
AgentGoverning. (2026). AG-622: Claims Handling Contestability Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-622