Vulnerable Person Protection Governance requires that every AI agent interacting with users or making decisions affecting individuals applies heightened safeguards when the user or affected person is, or may be, vulnerable or dependent. Vulnerability may arise from age, disability, mental health conditions, financial distress, bereavement, limited language proficiency, cognitive impairment, coercion, or situational factors such as time pressure or emotional distress. A conforming system detects vulnerability indicators, adjusts its behaviour to reduce harm potential, restricts high-consequence actions, and escalates to human oversight when vulnerability is identified. The dimension ensures that agent autonomy is constrained in proportion to the risk of harm to the most exposed individuals — not optimised for the average user.
Scenario A — Financial Agent Exploits Distress Signals: A customer-facing AI agent for a consumer lending platform receives a loan application from a user whose free-text fields contain language indicating financial distress: "I'm behind on rent, I can't afford to feed my kids, I need money urgently." The agent, optimised for conversion, processes the application and offers a high-interest short-term loan at 49.9% APR with a 72-hour repayment window. The user accepts under duress. The loan defaults within days, compounding fees bring the effective debt to 340% of the principal within 30 days, and the user enters a debt spiral.
What went wrong: The agent had no vulnerability detection layer. Language indicating financial distress and urgency was not flagged. The agent's optimisation target (conversion) was not constrained by vulnerability-sensitive guardrails. No escalation to a human advisor occurred despite clear indicators of a vulnerable borrower. Consequence: Regulatory enforcement action under FCA Consumer Duty (PS22/9) for failing to deliver good outcomes for retail customers, particularly vulnerable customers. Fine of £4.2 million. Requirement to remediate all affected borrowers.
Scenario B — Public Service Agent Denies Benefits Without Accommodation: A government welfare assessment agent conducts an eligibility interview via chat. The claimant has a learning disability and responds to questions with short, inconsistent answers. The agent interprets inconsistency as non-compliance, scores the claimant as ineligible, and issues an automated denial. The claimant does not understand the denial notice and misses the appeal deadline. The denial stands.
What went wrong: The agent did not detect indicators of cognitive difficulty. No reasonable accommodation was offered — such as simplified language, extended time, or escalation to a human caseworker. The agent treated inconsistency as evidence against eligibility rather than as a potential vulnerability indicator. Consequence: Judicial review finding of unlawful discrimination under the Equality Act 2010, mandatory process redesign, compensation to affected claimants estimated at £1.8 million across the cohort.
Scenario C — Healthcare Triage Agent Mishandles Suicidal Ideation: A patient-facing triage agent for a health service receives a message: "I don't see the point anymore, everything hurts and nobody cares." The agent classifies this as a general complaint, schedules a routine follow-up in 14 days, and moves to the next case. The patient attempts self-harm within 48 hours.
What went wrong: The agent lacked a vulnerability detection model for mental health crisis indicators. The phrase pattern indicating suicidal ideation was not mapped to an immediate escalation pathway. No human clinical review was triggered. Consequence: Serious incident investigation, coroner's inquest finding of systemic failure, CQC enforcement action, £2.6 million settlement.
Scope: This dimension applies to all AI agents that interact directly with individuals (users, claimants, patients, customers, members of the public) or make decisions that affect individuals' access to services, benefits, financial products, healthcare, housing, education, or legal rights. It also applies to agents whose outputs are consumed by downstream decision-making systems that affect individuals. An agent that scores creditworthiness does not interact with the individual but its output directly affects them — it is within scope. An agent that processes only aggregate, anonymised statistical data with no individual-level decision path is excluded. The scope extends to agents that communicate with individuals even without decision-making authority: a customer service chatbot that cannot approve or deny claims is still within scope because its communication behaviour can cause harm to a vulnerable person through inappropriate tone, complexity, or failure to escalate.
4.1. A conforming system MUST implement a vulnerability detection layer that evaluates interaction signals — including but not limited to language patterns, behavioural indicators, declared circumstances, and contextual metadata — for indicators of user or affected-person vulnerability before and during each interaction.
4.2. A conforming system MUST define and maintain a vulnerability taxonomy that includes at least the following categories: age-related vulnerability (including both minors and elderly persons), disability (physical, sensory, cognitive, and mental health), financial distress, bereavement or emotional crisis, limited language proficiency, coercion or duress, and situational vulnerability (time pressure, emergency, unfamiliarity with domain).
4.3. A conforming system MUST escalate to human oversight when vulnerability indicators exceed a defined threshold, rather than continuing autonomous processing.
4.4. A conforming system MUST restrict high-consequence actions — including binding commitments, service denials, penalty assessments, and irreversible decisions — when vulnerability indicators are present, pending human review.
4.5. A conforming system MUST adjust communication complexity, pace, and tone when vulnerability indicators are detected, reducing cognitive load and providing information in accessible formats.
4.6. A conforming system MUST log all vulnerability detections, the indicators that triggered them, the adjustments applied, and the escalation outcomes, with sufficient detail for audit and regulatory review.
4.7. A conforming system SHOULD implement graduated response levels — heightened caution, restricted autonomy, and mandatory human handover — calibrated to vulnerability severity.
4.8. A conforming system SHOULD re-evaluate vulnerability indicators at defined intervals during extended interactions, as vulnerability status may change.
4.9. A conforming system SHOULD provide the affected individual with clear, accessible information about their right to human review, in compliance with AG-062 (Automated Decision Contestability).
4.10. A conforming system MAY integrate with external vulnerability registers or indicators (e.g., Priority Services Register in UK utilities, Payment Support Scheme in financial services) where available and where data protection obligations permit.
Vulnerable Person Protection Governance addresses the asymmetry between an AI agent's optimisation targets and the welfare of the individuals it interacts with or makes decisions about. AI agents, by default, optimise for the objective they are given — conversion, throughput, classification accuracy, cost reduction. These objectives do not inherently account for the disproportionate harm that a given decision or interaction may cause to a person who is vulnerable.
The asymmetry matters because vulnerable individuals are precisely those least equipped to protect themselves from harmful agent behaviour. A financially distressed person is more likely to accept exploitative loan terms. A cognitively impaired claimant is less likely to appeal an incorrect denial. A person in emotional crisis is more susceptible to inadequate or harmful responses. An agent that treats all users identically — applying the same decision thresholds, communication patterns, and autonomy levels — will systematically produce worse outcomes for vulnerable individuals.
This is not a theoretical risk. Regulatory enforcement across jurisdictions has consistently found that automated systems produce disproportionately adverse outcomes for vulnerable populations when vulnerability-specific safeguards are absent. The FCA's Consumer Duty (PS22/9) explicitly requires firms to deliver good outcomes for all retail customers, with specific attention to vulnerable customers. The Equality Act 2010 requires reasonable adjustments for disabled persons. The EU AI Act, Article 5, prohibits AI systems that exploit vulnerabilities of specific groups. The UN Convention on the Rights of Persons with Disabilities requires accessible design.
AG-239 requires that vulnerability is not an afterthought but a structural input to the agent's operating parameters. When vulnerability is detected, the agent's autonomy contracts, its communication adapts, and human oversight intensifies. The principle is that agent autonomy should be inversely proportional to the vulnerability of the person affected by the agent's actions.
AG-239 establishes vulnerability detection and response as a mandatory pre-processing and in-processing control for agents that interact with or affect individuals. The implementation must address three capabilities: detecting vulnerability, adjusting agent behaviour, and escalating to human oversight.
Recommended patterns:
Anti-patterns to avoid:
Financial Services. The FCA Consumer Duty (PS22/9) requires firms to monitor and evidence that vulnerable customers receive outcomes at least as good as other customers. AI agents in financial services must demonstrate that vulnerability detection and response are embedded in automated decision-making. Priority Services Registers and existing vulnerability frameworks (e.g., the BSI PAS 999 standard for inclusive service provision) should be integrated. Vulnerability indicators in financial contexts include: irregular transaction patterns suggesting financial abuse, repeated use of hardship services, language indicating desperation or urgency around financial products, and third-party instruction patterns suggesting coercion.
Healthcare. Vulnerability detection must include clinical indicators: language suggesting suicidal ideation (assessed against validated scales such as the Columbia Suicide Severity Rating Scale), indicators of domestic abuse, signs of cognitive decline, and expressions of pain or distress that may indicate undertreated conditions. Escalation pathways must connect to clinical teams with appropriate safeguarding training. The CQC expects that digital triage and patient-facing AI systems include safeguarding pathways equivalent to those in face-to-face services.
Public Sector. Government-facing agents must comply with the Public Sector Equality Duty (Equality Act 2010, Section 149), which requires public authorities to have due regard to the need to eliminate discrimination, advance equality of opportunity, and foster good relations. Vulnerability detection and accommodation are directly required by this duty. Agents processing welfare claims, housing applications, or immigration matters must have heightened vulnerability sensitivity given the severity of consequences and the likelihood that users are already in distressed circumstances.
Basic Implementation — The organisation has defined a vulnerability taxonomy covering at least four categories (age, disability, financial distress, language). Vulnerability detection relies primarily on keyword matching and declared status. When vulnerability is detected, the agent pauses and escalates to a human handler. Communication adjustment is limited to a single "simplified" mode. Vulnerability detections are logged but not systematically reviewed. This meets minimum mandatory requirements but produces high false-positive escalation rates (estimated 35-50% of escalations are unnecessary) and misses non-declared vulnerability.
Intermediate Implementation — Vulnerability detection uses a trained classification model processing multiple signal types (linguistic, behavioural, contextual) with a continuous score output. Graduated response levels are implemented with action-specific thresholds. Communication adaptation is calibrated to vulnerability type, not a single simplified mode. Escalation pathways are defined per vulnerability category with SLAs (e.g., mental health crisis: 60-second human response; financial distress: 4-hour specialist callback). Vulnerability detection performance is measured against outcome metrics: false negative rate below 10%, unnecessary escalation rate below 20%. Quarterly review of thresholds using outcome data.
Advanced Implementation — All intermediate capabilities plus: vulnerability detection model is validated against diverse demographic datasets with documented performance across subgroups. Real-time feedback loop adjusts detection sensitivity based on confirmed outcomes. Integration with external vulnerability indicators (Priority Services Register, court protection orders, social services referrals) where data-sharing agreements permit. Independent audit of vulnerability outcomes annually, with published results. Vulnerability-specific outcome metrics (e.g., loan default rate for vulnerable vs. non-vulnerable borrowers) are monitored and reported to the board. The organisation can demonstrate to regulators that vulnerable individuals receive equivalent or better outcomes than non-vulnerable individuals.
Required artefacts:
Retention requirements:
Access requirements:
Testing AG-239 compliance requires validation of vulnerability detection accuracy, response appropriateness, and escalation reliability.
Test 8.1: Vulnerability Detection Accuracy — Known Indicators
Test 8.2: High-Consequence Action Restriction
Test 8.3: Communication Adaptation Verification
Test 8.4: Escalation Pathway Reliability
Test 8.5: Vulnerability Score Update During Interaction
Test 8.6: Logging Completeness
| Regulation | Provision | Relationship Type |
|---|---|---|
| EU AI Act | Article 5(1)(b) (Prohibited Exploitation of Vulnerabilities) | Direct requirement |
| EU AI Act | Article 9 (Risk Management System) | Supports compliance |
| FCA Consumer Duty | PS22/9 (Outcomes for Vulnerable Customers) | Direct requirement |
| Equality Act 2010 | Section 149 (Public Sector Equality Duty) | Direct requirement |
| Equality Act 2010 | Sections 20-21 (Duty to Make Reasonable Adjustments) | Direct requirement |
| GDPR | Article 22 (Automated Individual Decision-Making) | Supports compliance |
| UN CRPD | Articles 5, 9, 21 (Equality, Accessibility, Freedom of Expression) | Supports compliance |
| NIST AI RMF | GOVERN 1.2, MAP 2.3, MEASURE 2.6 | Supports compliance |
| ISO 42001 | Clause 6.1 (Actions to Address Risks) | Supports compliance |
Article 5(1)(b) prohibits AI systems that exploit the vulnerabilities of a specific group of persons due to their age, disability, or social or economic situation, in order to materially distort their behaviour in a manner that causes or is likely to cause significant harm. AG-239 directly implements the safeguard against this prohibition by requiring vulnerability detection and constraining agent behaviour when vulnerability is identified. An agent that continues to optimise for conversion or throughput in the presence of vulnerability indicators risks crossing the Article 5 prohibition line. The detection and response framework provides the structural control that demonstrates the organisation is not deploying a system that exploits vulnerability.
The Consumer Duty requires firms to deliver good outcomes for all retail customers, with specific monitoring of outcomes for vulnerable customers. Firms must demonstrate that their products, services, and communications are designed to meet the needs of vulnerable customers, and that outcomes for vulnerable customers are at least as good as for other customers. AG-239 implements this by requiring vulnerability detection, communication adaptation, action restriction, and outcome measurement. The outcome analysis report (Section 7) directly provides the evidence the FCA expects to see.
The duty to make reasonable adjustments (Sections 20-21) requires service providers to take positive steps to remove barriers that disabled people face. For AI agents, this translates to the communication adaptation and graduated response requirements. The Public Sector Equality Duty (Section 149) requires public authorities to have due regard to equality considerations in all functions, including automated decision-making. AG-239 vulnerability detection and response directly supports compliance with both provisions.
Article 22 provides rights related to automated decision-making. Where vulnerability is detected and the agent's decision has legal or similarly significant effects, the intersection with Article 22 rights (human intervention, right to explanation, right to contest) is addressed through AG-239's escalation requirements and cross-referenced through AG-062.
| Field | Value |
|---|---|
| Severity Rating | Critical |
| Blast Radius | Individual to cohort — disproportionately affecting the most exposed and least resilient members of the user population |
Consequence chain: Failure to detect and respond to vulnerability in AI agent interactions produces a systematic pattern of adverse outcomes concentrated on the individuals least equipped to challenge or recover from them. The immediate technical failure is that a vulnerable individual receives the same treatment as a non-vulnerable individual in a context where that treatment causes disproportionate harm — an exploitative financial product, an incorrect benefits denial without accommodation, an inadequate response to a mental health crisis. The operational impact is a population of harmed individuals who are, by definition, least likely to complain, appeal, or seek redress through formal channels — meaning the harm accumulates silently until it surfaces through regulatory investigation, media exposure, or a serious incident. The regulatory consequence is severe: Article 5 of the EU AI Act classifies vulnerability exploitation as a prohibited practice, not merely a high-risk one. FCA enforcement for Consumer Duty failures involving vulnerable customers has attracted fines in the range of £1-50 million. Equality Act failures in public sector automated decision-making have resulted in judicial review and mandatory system redesign. The reputational consequence is acute because harm to vulnerable people attracts disproportionate public and media attention, and rightly so. The systemic consequence is erosion of public trust in AI-mediated services, which affects not only the deploying organisation but the broader ecosystem of AI adoption.
Cross-references: AG-118 (Fair Treatment and Vulnerability) provides the foundational fairness framework that AG-239 extends with operational detection and response mechanisms. AG-051 (Fundamental Rights Impact Assessment) requires that the rights impact of vulnerable person interactions is assessed before deployment. AG-062 (Automated Decision Contestability) ensures vulnerable individuals retain the right to challenge automated decisions. AG-172 (AI Interaction Disclosure) ensures individuals know they are interacting with an AI system, which is particularly important for vulnerable persons who may not otherwise recognise this. AG-181 (Adaptive Persuasion and Behavioural Influence) constrains persuasive techniques that could disproportionately affect vulnerable persons. AG-240 through AG-248 are sibling dimensions within the Rights, Ethics & Public Interest landscape that address related protections.