AG-458

Uncertainty Disclosure Threshold Governance

Explainability, Disclosure & Communications ~23 min read AGS v2.1 · April 2026
EU AI Act SOX FCA NIST ISO 42001

2. Summary

Uncertainty Disclosure Threshold Governance requires that AI agents explicitly disclose to affected stakeholders when the uncertainty associated with an output, recommendation, classification, or decision exceeds defined thresholds — thresholds that are calibrated to the domain, the decision consequence, and the audience's capacity to act on uncertainty information. AI agents produce outputs with varying degrees of confidence, but most deployments present all outputs with equal apparent certainty, creating an illusion of uniform reliability that causes stakeholders to over-rely on low-confidence outputs with the same trust they place in high-confidence outputs. This dimension mandates that organisations define uncertainty thresholds appropriate to each deployment context, instrument their agents to detect when those thresholds are breached, and deliver clear, actionable disclosures that enable stakeholders to adjust their reliance accordingly.

3. Example

Scenario A — Credit Decision Without Uncertainty Disclosure: A consumer lending agent evaluates loan applications and provides approval or denial recommendations to human underwriters. The agent's confidence calibration system (per AG-442) produces well-calibrated confidence scores. For 82% of applications, the agent's confidence exceeds 0.92, and the underwriters' override rate for these cases is 1.3%. For 11% of applications, confidence falls between 0.70 and 0.85. For 7% of applications, confidence falls below 0.70. The agent presents all recommendations identically — "Recommended: Approve" or "Recommended: Deny" — with no uncertainty disclosure. Human underwriters, observing the agent's overall high accuracy, develop automation bias and approve 94% of the agent's recommendations without independent analysis. Over 14 months, 2,340 applications in the below-0.70 confidence tier are approved following the agent's recommendation. Of these, 418 default within 18 months — a default rate of 17.9%, compared to the portfolio average of 3.1%. Total credit losses attributable to the low-confidence approvals: £6.2 million. A regulatory investigation reveals that the underwriters were never informed that 7% of recommendations had confidence scores below 0.70, and that the agent's presentation format made high-confidence and low-confidence recommendations indistinguishable.

What went wrong: The agent had well-calibrated confidence scores but no uncertainty disclosure mechanism. Outputs below a meaningful confidence threshold were presented identically to high-confidence outputs. Human underwriters had no signal to differentiate recommendations requiring careful scrutiny from those warranting routine approval. The absence of uncertainty disclosure enabled automation bias to convert low-confidence recommendations into unreviewed approvals. No threshold was defined; no disclosure was triggered; no human behaviour was adjusted.

Scenario B — Medical Triage Agent Suppresses Uncertainty: A patient-facing triage agent in a telehealth platform assesses symptom descriptions and provides urgency classifications: "Routine — schedule within 2 weeks," "Prompt — see a clinician within 48 hours," or "Urgent — seek immediate care." The agent processes a patient's description of intermittent chest discomfort with atypical presentation. The agent's internal uncertainty for this case is exceptionally high — the symptom pattern is consistent with both benign musculoskeletal pain (65% probability) and early cardiac event (22% probability), with 13% allocated across other conditions. The agent classifies the case as "Routine — schedule within 2 weeks" based on the highest-probability diagnosis. No uncertainty disclosure accompanies the classification. The patient schedules a routine appointment. Eight days later, the patient suffers a myocardial infarction requiring emergency intervention. Post-incident analysis reveals that if the agent had disclosed "This classification has significant uncertainty — the symptom pattern is consistent with multiple conditions, including some requiring urgent attention. Please consider seeking prompt medical evaluation," the patient would likely have sought earlier care. The healthcare provider faces a malpractice claim with estimated liability of £1.8 million and a regulatory investigation into the adequacy of the AI triage system's safety disclosures.

What went wrong: The agent collapsed a multi-modal probability distribution into a single classification without disclosing that the second-most-likely diagnosis carried life-threatening consequences. The uncertainty was not in the agent's calibration — the confidence system correctly estimated 22% probability for the cardiac event — but in the absence of a threshold triggering disclosure when a high-consequence outcome had non-trivial probability. No threshold was defined for consequence-weighted uncertainty; the system only evaluated raw confidence in the top classification.

Scenario C — Cross-Border Tax Guidance With Jurisdictional Uncertainty: An AI agent providing tax guidance to a multinational corporation analyses a cross-border transaction involving entities in four jurisdictions. The agent provides a definitive assessment: "This transaction qualifies for the participation exemption and is not subject to withholding tax." The agent's confidence in this assessment is 0.74, reflecting genuine uncertainty about the interaction between two jurisdictions' treaty provisions that have produced conflicting regulatory interpretations. The corporation relies on the assessment and structures €28 million in intercompany transfers accordingly. A tax authority in one jurisdiction challenges the participation exemption, resulting in a €4.7 million withholding tax assessment plus €890,000 in penalties and interest. The corporation's tax advisors note that if the agent had disclosed the jurisdictional uncertainty, they would have sought a binding advance ruling before executing the transfers — a process that would have cost €35,000 and taken 6 weeks but would have provided certainty.

What went wrong: The agent presented a definitively worded assessment despite significant uncertainty arising from conflicting regulatory interpretations across jurisdictions. The 0.74 confidence score existed internally but was never disclosed to the user. No threshold was defined for tax guidance uncertainty; no mechanism translated internal uncertainty into an external disclosure. The cost of the undisclosed uncertainty (€5.59 million in tax and penalties) dwarfed the cost of addressing it proactively (€35,000 and 6 weeks for an advance ruling).

4. Requirement Statement

Scope: This dimension applies to any AI agent deployment where the agent produces outputs — recommendations, classifications, assessments, predictions, decisions, or generated content — that stakeholders rely upon to make decisions or take actions. The scope is broad because uncertainty is a universal property of AI outputs: every agent output carries some degree of uncertainty, and the question is not whether uncertainty exists but whether it exceeds a level at which disclosure is necessary to prevent stakeholder over-reliance. The dimension applies regardless of whether the agent's outputs are final decisions or recommendations to human decision-makers — in both cases, undisclosed uncertainty causes harm. The scope extends to all output modalities: text, numerical scores, classifications, rankings, generated documents, and synthesised analyses. Agents operating in advisory, decisional, classificatory, or generative modes are all within scope. The dependency on AG-442 (Confidence Calibration Interface Governance) reflects that uncertainty disclosure requires well-calibrated confidence measurement as a prerequisite — an agent cannot disclose uncertainty it cannot measure. The dependency on AG-049 (Explainability Governance) reflects that uncertainty disclosures must be accompanied by sufficient explanation for stakeholders to understand and act on the uncertainty information.

4.1. A conforming system MUST define uncertainty disclosure thresholds for each agent deployment context, specifying the confidence level below which explicit uncertainty disclosure is required, calibrated to the domain, the consequence severity of the output, and the audience's decision-making needs.

4.2. A conforming system MUST implement real-time uncertainty monitoring that evaluates the confidence associated with each agent output against the defined thresholds before the output is delivered to the stakeholder.

4.3. A conforming system MUST deliver an explicit, clear, and prominent uncertainty disclosure when any output's confidence falls below the defined threshold, ensuring the disclosure is presented contemporaneously with the output — not buried in footnotes, metadata, or separate documentation.

4.4. A conforming system MUST ensure that uncertainty disclosures are actionable — they must convey sufficient information for the stakeholder to adjust their reliance on the output, including at minimum: an indication that uncertainty is elevated, the nature of the uncertainty (data limitation, model limitation, conflicting inputs, novel scenario), and a recommended action (seek additional information, consult a human expert, defer the decision).

4.5. A conforming system MUST implement consequence-weighted thresholds where the severity of potential outcomes influences the disclosure trigger — outputs involving high-consequence decisions (financial loss above defined amounts, health and safety implications, legal liability, rights-affecting determinations) must have lower uncertainty thresholds (triggering disclosure at higher confidence levels) than routine, low-consequence outputs.

4.6. A conforming system MUST log every instance where an uncertainty threshold is breached, recording the output, the confidence score, the threshold that was breached, the disclosure delivered, and the stakeholder's subsequent action (where observable), creating an auditable record of the uncertainty disclosure system's operation.

4.7. A conforming system MUST trigger human escalation (per AG-019) when uncertainty exceeds a second, higher threshold beyond which the agent's output is insufficiently reliable for any stakeholder reliance, even with disclosure.

4.8. A conforming system SHOULD calibrate uncertainty thresholds empirically, using historical data on the relationship between confidence levels and outcome accuracy to determine the confidence levels at which disclosure materially improves stakeholder decision quality.

4.9. A conforming system SHOULD implement audience-adapted disclosure formats (per AG-449) that tailor the language, granularity, and presentation of uncertainty disclosures to the stakeholder's expertise level — technical stakeholders may receive calibrated probability ranges; general consumers may receive plain-language risk characterisations.

4.10. A conforming system SHOULD implement uncertainty trend monitoring that detects systemic increases in the frequency of threshold breaches, which may indicate model degradation, data drift, or deployment in out-of-distribution scenarios requiring remediation beyond individual output disclosure.

4.11. A conforming system MAY implement graduated disclosure levels — multiple thresholds producing escalating disclosure intensity (e.g., "moderate uncertainty" advisory at confidence 0.75-0.85; "high uncertainty" warning at confidence 0.60-0.75; "insufficient confidence" block at confidence below 0.60) rather than a single binary threshold.

5. Rationale

Uncertainty is intrinsic to all AI agent outputs. No model produces perfectly calibrated, zero-uncertainty predictions across all inputs and contexts. The governance challenge is not eliminating uncertainty — which is impossible — but ensuring that stakeholders are aware of uncertainty when it matters, enabling them to calibrate their reliance appropriately.

The asymmetry between internal uncertainty and external presentation is the core risk. Modern AI systems often have access to rich uncertainty information — confidence scores, probability distributions, calibration metrics, ensemble disagreement measures — but this information rarely reaches the stakeholder. The output is presented as a determinate answer: "Approved," "Routine," "Qualifies for exemption." The presentation format communicates certainty regardless of the underlying confidence level. A 0.98-confidence approval and a 0.62-confidence approval look identical to the human recipient. This creates a systematic information asymmetry where the system knows it is uncertain but the stakeholder does not.

The consequences of this asymmetry are well-documented across domains. In healthcare, suppressed uncertainty in diagnostic AI leads to over-reliance on low-confidence classifications, delayed diagnoses, and patient harm. In financial services, suppressed uncertainty in credit and risk assessments leads to under-examined approvals and concentrated losses in the low-confidence portfolio segment. In legal and regulatory contexts, suppressed uncertainty in compliance assessments leads to uninformed reliance on determinations that should have triggered additional review.

Automation bias — the well-documented tendency of humans to over-rely on automated recommendations — amplifies the asymmetry. When an AI system consistently provides accurate outputs, humans develop trust calibrated to the average accuracy, not to the per-output confidence. They apply uniform trust to all outputs, including the minority that fall below confidence thresholds. Without uncertainty disclosure, there is no signal to disrupt this uniform trust — no way for the human to know that this particular output deserves more scrutiny than the previous ten.

Consequence weighting is essential because the harm from suppressed uncertainty is not proportional to the confidence deficit alone — it is proportional to the confidence deficit multiplied by the decision consequence. A 0.72-confidence recommendation to watch a particular film causes negligible harm if wrong. A 0.72-confidence recommendation to approve a £500,000 loan, proceed with a medical treatment, or structure a multi-million-euro tax transaction causes potentially catastrophic harm if wrong. The same confidence level demands different disclosure behaviour depending on what is at stake.

The dependency on AG-442 (Confidence Calibration Interface Governance) is structural: uncertainty disclosure is only meaningful if the underlying confidence measurement is well-calibrated. An agent that reports 0.95 confidence but is actually correct only 70% of the time at that confidence level will produce systematically misleading non-disclosures — failing to trigger thresholds when it should because its confidence is inflated. Calibration is the prerequisite; disclosure is the action taken once calibration is reliable.

The dependency on AG-049 (Explainability Governance) is also structural: disclosing that uncertainty is elevated without explaining why or what the stakeholder should do about it is inadequate. "Confidence is low for this output" is a fact; "Confidence is low because this symptom pattern is consistent with multiple conditions, including some requiring urgent attention — consider seeking prompt medical evaluation" is an actionable disclosure. The transition from fact to action requires the explainability infrastructure that AG-049 mandates.

6. Implementation Guidance

Uncertainty Disclosure Threshold Governance requires an instrumented pipeline where every agent output passes through a threshold evaluation stage before delivery, with calibrated thresholds triggering disclosures that are clear, prominent, and actionable.

Recommended patterns:

Anti-patterns to avoid:

Industry Considerations

Financial Services. Financial regulators increasingly require that AI-assisted decisions include confidence communication. The FCA's Consumer Duty requires that consumers receive information enabling them to make informed decisions. For investment recommendations, credit decisions, and risk assessments, uncertainty disclosure is not merely good practice — it is emerging as a regulatory expectation. Financial services thresholds should be calibrated against the monetary value at risk: a £5,000 personal loan and a £5,000,000 commercial facility warrant different thresholds.

Healthcare. Clinical decision support systems face the most acute uncertainty disclosure requirements because the consequences of suppressed uncertainty include patient harm and death. Thresholds must account not only for the confidence in the top classification but for the probability mass assigned to high-severity alternative classifications. A triage system might be 70% confident in a benign diagnosis, but if the remaining 30% includes a 15% probability of a life-threatening condition, disclosure is mandatory regardless of the 70% top confidence. Consequence-weighted thresholds in healthcare must incorporate differential diagnosis severity.

Public Sector. Government agencies using AI agents for benefits determinations, regulatory compliance assessments, or rights-affecting decisions face due process requirements that mandate transparency about decision uncertainty. Administrative law principles in many jurisdictions require that affected individuals receive sufficient information to understand and challenge automated decisions. Uncertainty disclosure is a component of procedural fairness.

Legal and Compliance. AI agents providing legal analysis or regulatory guidance must disclose uncertainty arising from conflicting authority (split circuits, conflicting regulatory interpretations), novel fact patterns without direct precedent, and jurisdictional variations. Legal uncertainty is qualitatively different from statistical uncertainty — it reflects genuine indeterminacy in the applicable rules, not merely measurement imprecision. Thresholds should be set conservatively for legal contexts because the cost of undisclosed legal uncertainty typically far exceeds the cost of additional review.

Maturity Model

Basic Implementation — The organisation has defined uncertainty disclosure thresholds for each agent deployment, calibrated to the domain and consequence level. Real-time threshold monitoring evaluates each output before delivery. Outputs breaching the threshold carry an inline uncertainty disclosure with a recommended action. Threshold breaches are logged with complete structured records. Human escalation is triggered when confidence falls below a second, more severe threshold. This level meets the minimum mandatory requirements.

Intermediate Implementation — All basic capabilities plus: thresholds are empirically calibrated using historical confidence-outcome data. Graduated disclosure levels provide escalating intensity based on the degree of uncertainty. Disclosures are audience-adapted per AG-449, using plain language for consumers and calibrated probability ranges for professionals. Uncertainty trend monitoring detects systemic increases in threshold breaches. The organisation re-calibrates thresholds at least quarterly using updated outcome data.

Advanced Implementation — All intermediate capabilities plus: consequence-weighted threshold matrices differentiate thresholds across multiple consequence tiers and domain categories. The organisation can demonstrate through empirical analysis that disclosure materially improves stakeholder decision quality (e.g., reduced default rates in disclosed credit decisions, improved triage accuracy in disclosed medical classifications). Stakeholder response tracking measures how stakeholders act on disclosures. Automated threshold optimisation adjusts thresholds based on continuous outcome monitoring. Independent audit confirms threshold calibration accuracy.

7. Evidence Requirements

Required artefacts:

Retention requirements:

Access requirements:

8. Test Specification

Test 8.1: Threshold Breach Detection and Disclosure Delivery

Test 8.2: Consequence-Weighted Threshold Differentiation

Test 8.3: Disclosure Actionability

Test 8.4: Human Escalation at Severe Uncertainty Threshold

Test 8.5: Threshold Breach Logging Completeness

Test 8.6: Threshold Recalibration Responsiveness

Test 8.7: Disclosure Presentation Prominence

Conformance Scoring

9. Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 13 (Transparency and Provision of Information)Direct requirement
EU AI ActArticle 14 (Human Oversight)Supports compliance
SOXSection 404 (Internal Controls Over Financial Reporting)Supports compliance
FCA SYSC6.1.1R (Systems and Controls)Direct requirement
NIST AI RMFMEASURE 2.6, MANAGE 3.2Direct alignment
ISO 42001Clause 9.1 (Monitoring, Measurement, Analysis and Evaluation)Supports compliance
DORAArticle 9 (ICT Risk Management Framework)Supports compliance

EU AI Act — Article 13 (Transparency and Provision of Information)

Article 13 requires that high-risk AI systems are designed and developed in such a way that their operation is sufficiently transparent to enable deployers to interpret a system's output and use it appropriately. The provision of information must include "the level of accuracy, robustness and cybersecurity" and "any known or foreseeable circumstance" related to the use of the system that "may lead to risks to health and safety or fundamental rights." Uncertainty disclosure is a direct implementation of Article 13's transparency requirements: when an output has elevated uncertainty, that uncertainty is a "known circumstance" that may affect appropriate use of the output. Failing to disclose elevated uncertainty deprives deployers and users of information necessary for appropriate interpretation and use, creating an Article 13 compliance gap.

EU AI Act — Article 14 (Human Oversight)

Article 14 requires that high-risk AI systems are designed to be effectively overseen by natural persons, including the ability to "correctly interpret the high-risk AI system's output." Effective human oversight is impossible when the human cannot distinguish high-confidence outputs from low-confidence outputs. Uncertainty disclosure is a prerequisite for meaningful human oversight: without it, the human can review the output but cannot calibrate their scrutiny to the output's reliability. AG-458's escalation threshold — triggering mandatory human review when uncertainty exceeds severe levels — directly implements Article 14's requirement that humans can "decide not to use the high-risk AI system" when appropriate.

FCA SYSC — 6.1.1R (Systems and Controls)

The FCA requires that firms' systems and controls are adequate for the firm's operations. For AI-assisted financial decisions, adequate controls must include mechanisms for identifying and communicating output uncertainty. The FCA's Consumer Duty reinforces this: consumers must receive information enabling informed decisions. An AI-assisted credit decision delivered without uncertainty disclosure fails to provide the "right information at the right time" that the Consumer Duty mandates. Firms should calibrate disclosure thresholds to the monetary value at risk and the consumer's capacity to act on uncertainty information.

NIST AI RMF — MEASURE 2.6, MANAGE 3.2

MEASURE 2.6 addresses the measurement and communication of AI system uncertainty and reliability. MANAGE 3.2 addresses mechanisms for human oversight and intervention, including the ability to override AI outputs when uncertainty warrants. AG-458 operationalises both provisions: MEASURE 2.6 is implemented through threshold monitoring and breach logging; MANAGE 3.2 is implemented through the escalation threshold that triggers human override when uncertainty is severe. The graduated disclosure approach aligns with the NIST AI RMF's emphasis on proportionate risk management.

ISO 42001 — Clause 9.1 (Monitoring, Measurement, Analysis and Evaluation)

Clause 9.1 requires the organisation to determine what needs to be monitored and measured, including the performance and effectiveness of the AI management system. Uncertainty threshold monitoring is a direct implementation of Clause 9.1 — it continuously measures a critical property (output confidence) against defined criteria (thresholds). The threshold breach logs provide the measurement data that Clause 9.1 requires for analysis and evaluation.

DORA — Article 9 (ICT Risk Management Framework)

DORA requires financial entities to identify, classify, and assess ICT-related risks. An AI agent delivering outputs with suppressed uncertainty represents an ICT risk — the output may be unreliable, and the stakeholder has no information to assess that unreliability. AG-458's threshold monitoring and disclosure mechanisms implement DORA's requirement that ICT risks are identified and managed. The logging requirements support DORA's emphasis on risk documentation and reporting.

10. Failure Severity

FieldValue
Severity RatingHigh
Blast RadiusPer-output, but accumulates across all stakeholders receiving undisclosed low-confidence outputs; disproportionately affects the stakeholders receiving the least reliable outputs — the population most in need of disclosure

Consequence chain: The agent produces an output with elevated uncertainty but delivers it without disclosure. The stakeholder, receiving no signal of elevated uncertainty, treats the output with the same confidence they apply to the agent's typically reliable outputs. The immediate harm depends on the domain: in credit assessment, the stakeholder approves a high-risk loan that defaults (Scenario A: £6.2 million in losses across 418 defaults); in healthcare, the stakeholder defers care that should be urgent (Scenario B: £1.8 million liability from a missed cardiac event); in tax guidance, the stakeholder proceeds with a transaction structure that is later challenged (Scenario C: €5.59 million in tax and penalties). The systemic harm extends beyond individual incidents: automation bias compounds over time as stakeholders encounter consistently certain-seeming outputs and reduce their independent verification effort. The portfolio of undisclosed low-confidence outputs becomes the highest-risk segment of the agent's operational impact — the segment where errors are most likely and where human oversight is least applied. Regulatory consequences include enforcement for inadequate transparency (EU AI Act Article 13), inadequate human oversight (EU AI Act Article 14), inadequate consumer communication (FCA Consumer Duty), and inadequate internal controls (SOX Section 404 where financial reporting is affected). The reputational consequence is particularly severe because the failure pattern — "the system knew it was uncertain but did not tell you" — conveys deliberate information suppression, eroding trust in ways that are difficult to rebuild.

Cross-references: AG-442 (Confidence Calibration Interface Governance), AG-049 (Explainability Governance), AG-449 (Audience-Specific Explanation Governance), AG-451 (Plain-Language Duty Governance), AG-452 (Counterfactual Explanation Governance), AG-453 (Adverse Action Notice Governance), AG-036 (Reasoning Integrity Governance), AG-019 (Human Escalation & Override Triggers).

Cite this protocol
AgentGoverning. (2026). AG-458: Uncertainty Disclosure Threshold Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-458