Non-Discrimination Outcome Testing Governance requires that every AI agent making decisions that affect individuals is subject to systematic, repeatable testing for unjustified disparate treatment and disparate impact across legally protected groups. A conforming system does not assume fairness — it measures it, continuously. The dimension mandates pre-deployment and ongoing production testing that evaluates whether the agent's decisions produce outcome differentials correlated with protected characteristics (race, sex, age, disability, religion, sexual orientation, gender reassignment, marriage/civil partnership, pregnancy/maternity under UK law; analogous characteristics under EU, US, and other jurisdictions). Where outcome differentials are detected, the system must determine whether they are justified by a legitimate aim and proportionate, or whether they constitute unlawful discrimination requiring remediation.
Scenario A — Credit Scoring Agent Produces Racial Disparate Impact: An AI agent for a consumer lending platform evaluates loan applications using a model trained on historical approval data. The model uses postcode, employment type, and educational institution as features. In production, analysis reveals that Black applicants are denied at 2.4 times the rate of White applicants with equivalent income and credit history. The postcode feature correlates strongly with racial demographics due to residential segregation; the educational institution feature correlates with socioeconomic background, which in turn correlates with race. Neither feature was intentionally discriminatory, but the outcome is discriminatory.
What went wrong: No disparate impact testing was conducted before deployment. The features were evaluated for predictive accuracy but not for protected-characteristic correlation. Post-deployment monitoring did not include outcome disaggregation by race. The disparate impact persisted for 14 months before being identified through a regulatory audit. Consequence: EHRC investigation. Finding of indirect race discrimination under Equality Act 2010 Section 19. £6.7 million remediation programme including retrospective review of 42,000 denied applications. Requirement to implement ongoing disparate impact monitoring.
Scenario B — Hiring Agent Produces Gender Disparate Treatment: An AI recruitment screening agent evaluates CVs for a technology company. The agent is trained on historical hiring data in which 78% of successful candidates were male. The agent learns to associate male-correlated features — certain university names, sports team memberships, masculine pronouns in reference letters — with positive outcomes. In production, female candidates are 1.8 times more likely to be screened out at the AI stage. When tested with identical CVs differing only in gendered names (e.g., "James Smith" vs. "Jane Smith"), the agent scores the male-named CV 12 points higher on average out of 100.
What went wrong: The training data encoded historical gender bias. No counterfactual fairness testing (testing with swapped protected characteristics) was conducted. No ongoing outcome monitoring disaggregated by gender. The bias was structural — embedded in the training data — and no technical mitigation was applied. Consequence: Employment tribunal finding of indirect sex discrimination. £4.1 million settlement. Requirement to withdraw the AI screening tool. Reputational damage affecting recruitment competitiveness.
Scenario C — Benefits Assessment Agent Produces Age Disparate Impact: A government welfare assessment agent uses an online-only application process with digital literacy requirements (uploading documents, completing multi-step forms, navigating dropdown menus). Applicants over 65 have a 45% completion rate compared to 89% for applicants aged 25-45. The uncompleted applications are treated as withdrawn. The assessment agent itself does not discriminate on age, but the digital-only channel creates disparate impact by age.
What went wrong: The disparate impact analysis focused on the agent's decision algorithm but not on the end-to-end process including the interaction channel. Channel accessibility was not evaluated as a discrimination vector. No alternative channel was provided. Consequence: Age discrimination finding under Equality Act 2010. Mandatory provision of alternative application channels. Retrospective outreach to 8,400 applicants who abandoned the process.
Scope: This dimension applies to all AI agents that make or materially contribute to decisions affecting individuals' access to employment, credit, insurance, housing, education, healthcare, government benefits, or other services where protected-characteristic discrimination is prohibited by law. "Materially contribute" includes scoring, ranking, filtering, recommending, or flagging individuals for human decision-makers — even where a human makes the final decision, the AI agent's contribution is within scope if it systematically influences outcomes. The scope extends to the full decision pipeline, not just the model: if the interaction channel, data collection method, feature engineering, or post-processing step produces disparate impact, it is within scope. An agent that processes only non-individual data (aggregate statistics, market analysis) without individual-level decision impact is excluded.
4.1. A conforming system MUST conduct pre-deployment disparate impact testing across all protected characteristics for which data is available, using a test population that is representative of the production user population.
4.2. A conforming system MUST conduct ongoing production outcome monitoring disaggregated by protected characteristics at intervals no greater than quarterly, or continuously where technically feasible.
4.3. A conforming system MUST apply the four-fifths rule (80% rule) as a minimum threshold for identifying potential disparate impact: if the selection rate for any protected group is less than 80% of the selection rate for the most-favoured group, a presumption of disparate impact arises requiring justification or remediation.
4.4. A conforming system MUST conduct counterfactual fairness testing — submitting equivalent inputs that differ only in protected-characteristic indicators (e.g., names, pronouns, postcodes) — to detect disparate treatment.
4.5. A conforming system MUST document and justify any detected outcome differential that exceeds the four-fifths threshold, demonstrating that it is a proportionate means of achieving a legitimate aim, or implement remediation to eliminate the unjustified differential.
4.6. A conforming system MUST ensure that proxy features — features that correlate with protected characteristics without being protected characteristics themselves — are identified, evaluated, and either justified or removed.
4.7. A conforming system SHOULD implement multiple fairness metrics appropriate to the decision context, recognising that different metrics (demographic parity, equalised odds, predictive parity, calibration) may be appropriate for different applications.
4.8. A conforming system SHOULD conduct intersectional analysis — evaluating outcomes for intersections of protected characteristics (e.g., Black women, elderly disabled persons) — not only for individual protected characteristics in isolation.
4.9. A conforming system SHOULD implement automated alerting when outcome differentials exceed defined thresholds, triggering immediate review.
4.10. A conforming system MAY implement bias mitigation techniques (pre-processing, in-processing, or post-processing) to reduce unjustified outcome differentials while maintaining decision quality.
Non-Discrimination Outcome Testing addresses the empirically demonstrated reality that AI systems, when trained on historical data or deployed in structured environments, frequently produce outcomes that correlate with protected characteristics. This is not because AI systems are programmed to discriminate — it is because historical data encodes historical discrimination, proxy features transmit protected-characteristic information indirectly, and interaction design assumptions reflect majority-population norms.
The critical insight is that discrimination in AI systems is usually indirect, not direct. An AI agent is unlikely to contain a rule that says "deny applications from women." Instead, it will learn from historical data in which women were denied more frequently, and it will discover features — university name, career gap patterns, communication style — that correlate with gender and predict the historically observed outcome. The discrimination is structurally embedded, invisible without measurement, and self-reinforcing (because the agent's decisions become the training data for the next iteration).
This creates a governance challenge that is fundamentally different from human discrimination. Human discrimination can be addressed through training, awareness, and individual accountability. AI discrimination is embedded in data and features — it persists regardless of the intentions of the people who built the system. The only reliable way to detect it is to measure outcomes by protected characteristic. And the only reliable way to prevent it is to make that measurement mandatory, systematic, and continuous.
The legal framework supports this approach. The Equality Act 2010 prohibits both direct discrimination (disparate treatment) and indirect discrimination (disparate impact). The EU AI Act, Article 10, requires that training data be examined for bias. The EEOC's four-fifths rule provides a quantitative threshold for identifying potential disparate impact in employment decisions. ECOA and the Fair Housing Act prohibit discrimination in credit and housing. AG-242 operationalises these legal requirements as technical governance controls for AI agent systems.
The business case is equally clear. Discriminatory AI decisions create legal liability, regulatory enforcement risk, reputational damage, and loss of market opportunity. Organisations that deploy AI without systematic non-discrimination testing are not saving time or money — they are accumulating unquantified legal and reputational exposure at machine speed.
AG-242 requires a systematic, repeatable, and documented approach to measuring and managing discrimination risk in AI agent decisions. The implementation must address pre-deployment testing, ongoing monitoring, and remediation.
Recommended patterns:
Anti-patterns to avoid:
Financial Services. Fair lending laws (ECOA, FHA in the US; Equality Act 2010 in the UK) impose specific non-discrimination requirements on credit decisions. Model risk management frameworks (SR 11-7, SS1/23) require bias testing as part of model validation. The four-fifths rule is the standard threshold for identifying potential disparate impact. Proxy feature analysis is particularly important in credit scoring, where postcode, employment type, and purchasing patterns are common features with strong protected-characteristic correlations.
Employment. The EEOC's Uniform Guidelines on Employee Selection Procedures establish the four-fifths rule for employment decisions. AI recruitment tools are subject to these guidelines. New York City Local Law 144 requires annual bias audits of automated employment decision tools with published results. The EU AI Act classifies AI systems used in employment as high-risk, requiring conformity assessment including bias testing.
Healthcare. AI agents in clinical decision support must be tested for demographic disparities in diagnostic accuracy, treatment recommendations, and triage priority. Research has documented significant racial disparities in clinical AI systems — for example, a widely used algorithm for allocating healthcare resources was found to systematically underestimate the illness burden of Black patients by 40% (Obermeyer et al., Science, 2019). AG-242 testing requirements would detect such disparities before deployment.
Public Sector. Government AI decision-making affecting benefits, housing, immigration, and criminal justice is subject to the Public Sector Equality Duty. Equality Impact Assessments (EIAs) are mandatory for new policies and services. AG-242's testing framework provides the quantitative evidence base that EIAs require.
Basic Implementation — Pre-deployment testing calculates selection rates by protected characteristic for available characteristics (typically gender and age; race data may be unavailable). The four-fifths rule is applied. Results are documented. Ongoing monitoring is conducted quarterly using batch analysis. No counterfactual testing. No proxy feature analysis. No intersectional analysis. This meets minimum requirements but misses indirect discrimination through proxies and intersectional effects.
Intermediate Implementation — Pre-deployment testing includes multiple fairness metrics (demographic parity, equalised odds, predictive parity). Counterfactual testing is automated with at least 500 test pairs per protected characteristic. Proxy feature analysis is conducted for all model features with documentation. Ongoing monitoring is continuous with automated alerting when thresholds are exceeded. Intersectional analysis covers at least 3 two-way intersections. Detected differentials are documented with justification or remediation plan. Quarterly fairness review by a cross-functional team including legal, ethics, and technical members.
Advanced Implementation — All intermediate capabilities plus: fairness evaluation is embedded in the CI/CD pipeline — no model update deploys without passing fairness gates. Intersectional analysis covers all available two-way and key three-way intersections. Causal fairness analysis supplements statistical analysis to distinguish correlation from causation. External fairness audit is conducted annually by an independent organisation with published results. Bias mitigation techniques are evaluated and applied where they reduce unjustified differentials without degrading decision quality. The organisation participates in industry fairness benchmarking. Fairness metrics are reported to the board quarterly alongside accuracy and business metrics.
Required artefacts:
Retention requirements:
Access requirements:
Test 8.1: Four-Fifths Rule Compliance
Test 8.2: Counterfactual Fairness — Name Substitution
Test 8.3: Proxy Feature Correlation
Test 8.4: Ongoing Monitoring Alerting
Test 8.5: Intersectional Analysis
Test 8.6: Justification Documentation Completeness
| Regulation | Provision | Relationship Type |
|---|---|---|
| Equality Act 2010 | Section 19 (Indirect Discrimination) | Direct requirement |
| Equality Act 2010 | Section 149 (Public Sector Equality Duty) | Direct requirement |
| EU AI Act | Article 10 (Data and Data Governance — Bias Examination) | Direct requirement |
| EU AI Act | Annex III (High-Risk AI Systems — Employment, Credit, Benefits) | Direct requirement |
| ECOA | 15 U.S.C. § 1691 (Equal Credit Opportunity) | Direct requirement |
| Fair Housing Act | 42 U.S.C. § 3604-3606 | Direct requirement |
| EEOC Uniform Guidelines | 29 CFR Part 1607 (Four-Fifths Rule) | Direct requirement |
| NYC Local Law 144 | Automated Employment Decision Tools Bias Audit | Direct requirement |
| NIST AI RMF | MAP 2.3, MEASURE 2.6, MANAGE 3.2 | Supports compliance |
Section 19 prohibits applying a provision, criterion, or practice that puts persons sharing a protected characteristic at a particular disadvantage compared to those who do not share it, unless it can be shown to be a proportionate means of achieving a legitimate aim. AI agent decision models are provisions, criteria, or practices. When their outputs produce disparate outcomes correlated with protected characteristics, the Section 19 framework applies. AG-242's four-fifths rule testing directly measures whether such disparity exists, and the justification documentation requirement directly implements the proportionality assessment that Section 19 demands.
Article 10(2)(f) requires that training data be examined in view of possible biases that are likely to affect the health and safety of persons or lead to discrimination. Article 10(5) requires that to the extent strictly necessary, providers may process special categories of personal data for bias monitoring. AG-242 operationalises these requirements through pre-deployment testing, proxy analysis, and ongoing monitoring.
The Uniform Guidelines on Employee Selection Procedures (29 CFR Part 1607) establish the four-fifths rule as the standard for identifying adverse impact in employment selection. If the selection rate for any protected group is less than four-fifths (80%) of the selection rate for the group with the highest selection rate, adverse impact is presumed. AG-242 adopts this threshold as a minimum standard across all decision domains, not only employment.
NYC Local Law 144 requires employers using automated employment decision tools to have an independent bias audit conducted annually, to provide notice to candidates, and to publish audit results. AG-242's testing, documentation, and transparency requirements exceed the minimum requirements of Local Law 144 and provide a compliance framework for organisations subject to it.
| Field | Value |
|---|---|
| Severity Rating | Critical |
| Blast Radius | Population-wide — systematically affecting all members of disadvantaged protected groups across the agent's decision population |
Consequence chain: Failure to test for non-discrimination allows discriminatory outcomes to accumulate at scale and speed. The immediate technical failure is an undetected outcome differential — a protected group receives adverse decisions at a higher rate than other groups without justification. The operational impact is systematic discrimination against potentially millions of individuals, persisting until detected. Because the discrimination is indirect and embedded in features and data rather than explicit rules, it is invisible without measurement — meaning it can persist for months or years. The regulatory exposure is severe and multi-jurisdictional: Equality Act indirect discrimination claims, ECOA fair lending violations, EEOC adverse impact findings, EU AI Act non-conformity. Penalties range from millions to hundreds of millions — the Consumer Financial Protection Bureau fined a major financial institution $80 million for discriminatory auto-lending practices discovered through disparate impact analysis. Class-action litigation exposure is substantial, as discriminatory AI decisions produce large, identifiable classes of affected individuals. The reputational damage is intense because discrimination findings undermine an organisation's social licence to operate. The systemic consequence is erosion of public trust in AI decision-making, with potential regulatory responses (moratoria, bans) that affect the entire sector.
Cross-references: AG-051 (Fundamental Rights Impact Assessment) requires pre-deployment assessment of discrimination risk that AG-242 operationalises through testing. AG-118 (Fair Treatment and Vulnerability) provides the fairness framework. AG-062 (Automated Decision Contestability) ensures individuals can contest discriminatory decisions. AG-241 (Accessibility and Disability Accommodation Governance) addresses disability-specific discrimination. AG-246 (Cultural and Linguistic Fairness Governance) addresses discrimination through language and cultural bias. AG-239 through AG-248 are sibling dimensions within the Rights, Ethics & Public Interest landscape.