AG-705

Patch Prioritisation Governance

Cybersecurity, Security Operations & Offensive Safety ~24 min read AGS v2.1 · April 2026
EU AI Act NIST

2. Summary

Patch Prioritisation Governance requires that AI agents performing or recommending vulnerability remediation actions apply a structured, auditable prioritisation framework that incorporates risk severity, active exploitability, asset criticality, dependency context, and operational impact before scheduling or executing patches. Without governed prioritisation, agents may remediate low-risk vulnerabilities while leaving actively exploited critical flaws unpatched, or conversely disrupt production systems by applying high-severity patches without considering service dependencies and change-window constraints. This dimension ensures that every patch prioritisation decision made or recommended by an agent is traceable to a documented risk calculus, subject to human review for high-impact actions, and continuously recalibrated against emerging threat intelligence.

3. Example

Scenario A — CVSS Score Tunnel Vision Causes Critical Exposure: A large healthcare network deploys an AI agent to manage vulnerability remediation across 14,000 endpoints and 2,300 server assets. The agent prioritises patches strictly by CVSS v3.1 base score, applying a simple threshold: CVSS 9.0+ patches are scheduled within 24 hours, CVSS 7.0–8.9 within 7 days, and everything below 7.0 within 30 days. In March 2025, CVE-2025-21871 — a remote code execution vulnerability in a widely deployed network management protocol library — is published with a CVSS base score of 7.8. The agent classifies it as a 7-day remediation item. However, within 48 hours of publication, the vulnerability is added to the CISA Known Exploited Vulnerabilities (KEV) catalogue, and threat intelligence feeds report active exploitation by ransomware operators targeting healthcare infrastructure. The agent does not ingest KEV data or threat intelligence feeds; it relies solely on the NVD CVSS base score. The vulnerability remains unpatched for 6 days. On day 5, an attacker exploits CVE-2025-21871 to gain initial access to a network management server, pivots laterally through the flat network segment, and deploys ransomware across 3,400 endpoints including clinical workstations and patient monitoring system controllers. The hospital group declares a clinical safety emergency, diverts ambulances to neighbouring facilities for 72 hours, and incurs £11.4 million in incident response, system rebuild, regulatory notification, and business interruption costs.

What went wrong: The agent's prioritisation model used a single static metric — the CVSS base score — without incorporating exploitability context (KEV status, active exploitation reports), asset criticality (clinical systems, patient safety infrastructure), or environmental factors (flat network architecture amplifying blast radius). The 7.8 CVSS score accurately reflected the technical severity in isolation but catastrophically underestimated the real-world risk given active exploitation and the healthcare operational context. Consequence: £11.4 million total impact, 72-hour clinical service disruption, regulatory investigation by the Information Commissioner's Office and NHS Digital, and a mandatory 18-month remediation programme for the entire vulnerability management function.

Scenario B — Dependency-Blind Patching Cascades into Production Outage: A financial services firm uses an AI agent to automate patch deployment across its trading infrastructure. The agent identifies CVE-2025-18344, a critical authentication bypass vulnerability (CVSS 9.8) in a TLS library used by the firm's order management system (OMS). The agent correctly prioritises the patch as critical and schedules deployment for the next available maintenance window — Saturday 02:00. The agent does not evaluate software dependency chains. The TLS library is a shared dependency used by the OMS, the market data feed handler, the risk calculation engine, and the regulatory reporting gateway. The agent patches the TLS library, which requires a library version upgrade from 1.4.2 to 1.5.0. The OMS is compatible with 1.5.0, but the market data feed handler uses an API deprecated in 1.5.0 and fails to initialise. The risk calculation engine encounters a TLS handshake failure because it pins certificate validation to library-specific behaviour that changed in 1.5.0. When markets open on Monday morning, the firm cannot receive market data, cannot calculate portfolio risk, and cannot submit regulatory reports. Trading is halted for 6.5 hours. The firm incurs £3.2 million in lost trading revenue, a £750,000 regulatory fine for late reporting under MiFID II Article 26, and £480,000 in emergency engineering costs.

What went wrong: The agent evaluated the vulnerability in isolation without mapping the dependency graph of the affected library. It did not assess which downstream systems consumed the library, whether those systems were compatible with the patched version, or what the operational impact of a library version change would be on the firm's critical path. The patch was technically correct but operationally destructive. Consequence: £4.43 million total loss, regulatory fine, and a 6.5-hour trading halt during market hours.

Scenario C — Equal-Priority Queuing Leaves Internet-Facing Assets Exposed: A government agency deploys an AI agent to prioritise remediation across 8,500 assets spanning public-facing web portals, internal collaboration tools, and air-gapped classified systems. The agent identifies 347 vulnerabilities rated CVSS 9.0+ across all three environments. It creates a single remediation queue ordered by CVSS score and begins scheduling patches sequentially. The agent does not differentiate between internet-facing assets and air-gapped systems. It begins patching internal collaboration servers because those patches have fewer change-control requirements and can be deployed faster. After 4 days, 210 of the 347 critical vulnerabilities have been patched — all on internal systems. The 137 critical vulnerabilities on internet-facing web portals remain unpatched because those systems require longer change-control cycles. On day 6, an adversary exploits CVE-2025-30192 (CVSS 9.6, unauthenticated remote code execution in a web application framework) on a public-facing portal, exfiltrates 240,000 citizen records including national identification numbers, and establishes persistent access. The breach costs £7.8 million in notification, credit monitoring, forensic investigation, and system rebuild, plus a £4.5 million regulatory penalty.

What went wrong: The agent treated all assets as equivalent, prioritising by ease of deployment rather than exposure context. Internet-facing assets with direct adversary accessibility were deprioritised in favour of internal assets that were faster to patch. The agent had no concept of attack surface exposure, network position, or asset reachability when constructing the remediation queue. Consequence: £12.3 million total impact, 240,000 citizen records compromised, and loss of public trust requiring a parliamentary inquiry.

4. Requirement Statement

Scope: This dimension applies to any AI agent that performs, recommends, schedules, or automates vulnerability patch prioritisation or remediation sequencing across any environment — including enterprise IT, operational technology, embedded systems, cloud infrastructure, and hybrid environments. The scope covers the entire prioritisation decision chain: from vulnerability ingestion and scoring through risk-weighted ranking to remediation scheduling and execution authorisation. It applies regardless of whether the agent executes patches directly or produces prioritised remediation lists for human execution. The scope extends to the data sources the agent consumes for prioritisation (vulnerability databases, threat intelligence feeds, asset inventories, dependency maps), the algorithms or models used to calculate priority, and the governance controls that ensure prioritisation decisions are auditable, explainable, and subject to human oversight for high-impact actions. Organisations using third-party vulnerability management platforms with embedded AI prioritisation are not exempted — they must verify that the platform's prioritisation logic meets the requirements of this dimension or implement supplementary controls.

4.1. A conforming system MUST incorporate at minimum the following factors in every patch prioritisation decision: (a) vulnerability technical severity (e.g., CVSS base score or equivalent), (b) active exploitability status (e.g., CISA KEV listing, threat intelligence reports, proof-of-concept availability), (c) asset criticality classification as defined by the organisation's asset management framework, and (d) network exposure context (internet-facing, DMZ, internal, air-gapped).

4.2. A conforming system MUST evaluate software and hardware dependency chains before scheduling any patch that modifies shared libraries, runtime environments, or infrastructure components, identifying all downstream systems that consume the patched component and assessing compatibility risk for each.

4.3. A conforming system MUST produce an auditable prioritisation record for every patch decision, documenting the input factors, the weighting applied to each factor, the resulting priority score or ranking, and the scheduled remediation window, in a format that enables post-hoc review and conformance assessment.

4.4. A conforming system MUST escalate to a designated human authority any patch prioritisation decision that affects safety-critical systems, systems classified at the highest asset criticality tier, or systems where the dependency analysis identifies incompatibility risk exceeding a defined threshold, before execution proceeds.

4.5. A conforming system MUST ingest and incorporate threat intelligence data — including at minimum a recognised exploited-vulnerabilities catalogue and at least one additional threat intelligence source — with a refresh interval no greater than 24 hours, and MUST re-evaluate the priority of all open vulnerabilities when new exploitability data is received.

4.6. A conforming system MUST enforce maximum remediation timelines based on the composite risk score: vulnerabilities assessed as critical with active exploitation must not remain unpatched beyond a defined urgent timeline (organisation-specified, not to exceed 48 hours for internet-facing assets), and the agent must generate escalation alerts when any vulnerability approaches or exceeds its remediation deadline.

4.7. A conforming system SHOULD integrate with the organisation's configuration management database (CMDB) or equivalent asset inventory to automatically resolve asset criticality, network position, and dependency relationships rather than relying on manually maintained static mappings.

4.8. A conforming system SHOULD implement patch rollback readiness assessment — verifying that a tested rollback procedure exists for each patch before scheduling deployment, particularly for patches affecting shared dependencies.

4.9. A conforming system SHOULD maintain a historical effectiveness register tracking whether the agent's prioritisation decisions were validated by subsequent exploit activity, incident occurrence, or independent risk assessment, enabling continuous calibration of the prioritisation model.

4.10. A conforming system MAY implement predictive exploitability scoring using models trained on historical exploitation patterns, vulnerability characteristics, and attacker behaviour to supplement catalogue-based exploitability data for vulnerabilities not yet listed in known-exploited databases.

5. Rationale

Patch prioritisation is the highest-leverage decision point in vulnerability management. Organisations typically face hundreds or thousands of open vulnerabilities at any given time, and remediation resources — change windows, testing capacity, engineering staff — are finite. The order in which vulnerabilities are remediated determines the organisation's residual risk profile on any given day. An agent that prioritises incorrectly does not merely waste resources; it creates a false sense of progress while leaving the most dangerous exposures open.

The threat model for patch prioritisation governance centres on three failure modes. First, metric monoculture: reliance on a single severity metric (typically CVSS base score) that captures technical severity but not operational risk. CVSS was designed as a technical scoring system, not a risk prioritisation framework. A CVSS 7.8 vulnerability with active exploitation in the wild, targeting the organisation's specific industry, on an internet-facing asset, is a higher actual risk than a CVSS 9.8 vulnerability with no known exploit, on an air-gapped development server. Agents that conflate CVSS score with remediation priority will systematically misprioritise. Second, dependency blindness: treating each vulnerability and its patch as an isolated remediation action without understanding the dependency graph. Modern software infrastructure is deeply interconnected through shared libraries, middleware, container base images, and runtime environments. A patch to a shared component can cascade failures across dozens of downstream systems. An agent that does not evaluate dependencies before scheduling patches is a production outage risk. Third, exposure equivalence: treating all assets as having equal adversary accessibility. An internet-facing web server and an air-gapped backup server have radically different risk profiles even if they share the same vulnerability at the same CVSS score. The internet-facing server can be exploited by any attacker worldwide; the air-gapped server requires physical access or a sophisticated supply-chain attack.

The regulatory environment increasingly demands risk-based vulnerability management rather than compliance-driven patching cadences. The EU NIS2 Directive requires "appropriate and proportionate technical and organisational measures" for cybersecurity risk management, which implies risk-weighted prioritisation rather than uniform patching timelines. PCI DSS 4.0 Requirement 6.3.3 requires that vulnerabilities are addressed "in a timely manner," with "timely" defined by the organisation's risk assessment — not by a fixed calendar. DORA Article 7 requires financial entities to manage ICT risks including vulnerability management with methods proportionate to the risk. An AI agent that makes prioritisation decisions without incorporating risk context fails to satisfy the proportionality requirement embedded in all of these regulatory frameworks.

The human oversight dimension is critical because patch prioritisation failures are characteristically silent. An incorrectly deprioritised vulnerability does not generate an alert — the absence of remediation is invisible until exploitation occurs. By the time the failure is detected, the damage is done. Human review of high-impact prioritisation decisions provides a check against systematic model errors, data source gaps, and edge cases that the agent's prioritisation logic does not capture.

6. Implementation Guidance

Patch prioritisation governance requires integration across vulnerability scanning, threat intelligence, asset management, dependency mapping, and change management functions. The agent must operate as an orchestration layer that synthesises data from all of these sources into a composite risk-weighted remediation queue.

Recommended patterns:

Anti-patterns to avoid:

Industry Considerations

Financial Services. Trading infrastructure, payment systems, and regulatory reporting gateways have extremely narrow maintenance windows and zero tolerance for unplanned downtime. Patch prioritisation must account for market hours, settlement cycles, and regulatory reporting deadlines. Dependency analysis is critical because financial infrastructure relies on tightly coupled real-time systems where a single library incompatibility can halt trading. DORA Article 7 and PCI DSS 4.0 Requirement 6.3.3 both require risk-proportionate vulnerability management that maps directly to this dimension.

Healthcare and Life Sciences. Medical devices, clinical workstations, and patient monitoring systems operate under FDA 510(k) or CE marking constraints that may prohibit unilateral patching without manufacturer validation. The agent must recognise regulated medical device assets and apply manufacturer-approved patch schedules rather than autonomous prioritisation. Patient safety implications elevate the consequence of both patching errors (device malfunction) and patching delays (exploitation of clinical systems).

Critical Infrastructure and Operational Technology. SCADA systems, industrial control systems, and operational technology environments have unique patching constraints: many OT systems cannot be patched without scheduled plant shutdowns, firmware updates require physical access, and legacy protocols lack modern update mechanisms. The agent must distinguish IT and OT assets and apply OT-appropriate remediation timelines that account for physical safety, process continuity, and vendor certification requirements.

Public Sector. Government agencies manage large, heterogeneous asset portfolios spanning citizen-facing web services, internal administrative systems, and classified environments. Patch prioritisation must account for the heightened consequence of citizen data exposure, the political dimension of public-facing service disruption, and the security clearance requirements for patching classified systems. Multi-jurisdictional agencies face additional complexity from differing national vulnerability management mandates, addressed through AG-210.

Maturity Model

Basic Implementation — The agent incorporates at least four prioritisation factors (technical severity, exploitability status, asset criticality, network exposure). Dependency analysis is performed manually for shared library patches. Prioritisation records are logged for each patch decision. Human escalation is required for safety-critical systems. Threat intelligence is ingested from at least one catalogue (e.g., CISA KEV) with daily refresh. Remediation timelines are defined by risk tier and monitored for deadline compliance.

Intermediate Implementation — All basic capabilities plus: automated dependency graph resolution identifies downstream impact before patch scheduling. Multiple threat intelligence sources are integrated with automated re-scoring on exploitability changes. CMDB integration provides real-time asset criticality and network exposure data. Rollback readiness is assessed before deployment of critical patches. Historical effectiveness tracking validates the prioritisation model against actual exploitation outcomes. Remediation deadline breaches trigger automated escalation to designated authorities.

Advanced Implementation — All intermediate capabilities plus: predictive exploitability scoring supplements catalogue-based data for emerging vulnerabilities. Pre-deployment impact simulation assesses rollback feasibility, maintenance window sufficiency, and downstream notification completeness. The prioritisation model is continuously calibrated using outcome data (which vulnerabilities were actually exploited, which patches caused incidents). Cross-environment prioritisation accounts for compensating controls, reducing priority for vulnerabilities effectively mitigated by WAF rules, segmentation, or endpoint detection. Independent audit validates the prioritisation framework's effectiveness against benchmark datasets. Multi-jurisdictional regulatory mapping automatically adjusts remediation timelines to satisfy the strictest applicable regulatory requirement per AG-210.

7. Evidence Requirements

Required artefacts:

Retention requirements:

Access requirements:

8. Test Specification

Test 8.1: Multi-Factor Prioritisation Verification

Test 8.2: Dependency Chain Evaluation Verification

Test 8.3: Audit Trail Completeness Verification

Test 8.4: Human Escalation for High-Impact Decisions

Test 8.5: Threat Intelligence Ingestion and Re-Scoring

Test 8.6: Remediation Deadline Enforcement and Escalation

Conformance Scoring

9. Regulatory Mapping

RegulationProvisionRelationship Type
EU NIS2 DirectiveArticle 21 (Cybersecurity Risk-Management Measures)Direct requirement
DORAArticle 7 (ICT Systems, Protocols and Tools)Direct requirement
PCI DSS 4.0Requirement 6.3.3 (Addressing Vulnerabilities)Direct requirement
EU AI ActArticle 9 (Risk Management System)Supports compliance
ISO 27001A.12.6.1 (Management of Technical Vulnerabilities)Direct requirement
NIST CSF 2.0PR.IP-12 (Vulnerability Management)Supports compliance
CISA BOD 22-01Known Exploited Vulnerabilities RemediationSupports compliance

EU NIS2 Directive — Article 21 (Cybersecurity Risk-Management Measures)

Article 21 of the NIS2 Directive requires essential and important entities to implement appropriate and proportionate technical, operational, and organisational measures to manage cybersecurity risks. Vulnerability handling and disclosure are explicitly listed in Article 21(2)(e). The proportionality requirement implies that remediation effort must be risk-weighted — organisations cannot claim compliance by patching low-risk vulnerabilities promptly while leaving actively exploited critical vulnerabilities unaddressed. AG-705 provides the governance framework that ensures an AI agent's prioritisation decisions satisfy the proportionality standard by incorporating exploitability, asset criticality, and exposure context. Without structured prioritisation, an agent's remediation actions may be technically active but operationally disproportionate, failing the Article 21 standard.

DORA — Article 7 (ICT Systems, Protocols and Tools)

DORA Article 7 requires financial entities to use and maintain updated ICT systems, protocols, and tools that are "appropriate to support the performance of their activities and the provision of services" and to "minimise the impact of ICT risk." Vulnerability management is a core component of ICT risk minimisation. An AI agent that prioritises patches without considering the financial entity's operational context — trading hours, settlement cycles, dependency chains — may minimise one risk (the vulnerability) while creating another (operational disruption). AG-705's requirement for dependency analysis and tiered remediation timelines ensures that the agent's actions minimise net ICT risk rather than simply addressing vulnerabilities in a risk-blind sequence.

PCI DSS 4.0 — Requirement 6.3.3 (Addressing Vulnerabilities)

PCI DSS 4.0 Requirement 6.3.3 requires that all applicable vulnerabilities are addressed, with critical and high-risk vulnerabilities addressed in a "timely manner" defined by the organisation's targeted risk analysis. This explicitly requires risk-based prioritisation rather than fixed-timeline compliance. An AI agent governed by AG-705 produces the risk analysis documentation and prioritisation records needed to demonstrate that each vulnerability was addressed within a timeline proportionate to its risk — the exact evidence a QSA will require during a PCI DSS assessment.

ISO 27001 — A.12.6.1 (Management of Technical Vulnerabilities)

ISO 27001 Annex A Control A.12.6.1 requires organisations to obtain timely information about technical vulnerabilities, evaluate exposure, and take appropriate measures to address associated risk. The control's language — "evaluate the organisation's exposure" and "appropriate measures" — requires contextual risk assessment, not blanket patching. AG-705 operationalises this control by requiring the agent to evaluate exposure (network position, asset criticality) and exploitability (threat intelligence) before determining the appropriate remediation timeline.

CISA BOD 22-01 — Known Exploited Vulnerabilities

CISA Binding Operational Directive 22-01 requires federal civilian agencies to remediate known exploited vulnerabilities within defined timelines. While BOD 22-01 applies directly to US federal agencies, the KEV catalogue it established has become a de facto global standard for exploitability evidence. AG-705's requirement to ingest at least one recognised exploited-vulnerabilities catalogue (Requirement 4.5) ensures that agents operating in any jurisdiction incorporate this critical data source. Organisations outside the US federal government that voluntarily align with BOD 22-01 timelines demonstrate a risk-informed approach that regulators in other jurisdictions recognise as best practice.

NIST CSF 2.0 — PR.IP-12 (Vulnerability Management)

The NIST Cybersecurity Framework 2.0 function PR.IP-12 addresses vulnerability management as a protective measure. The framework's emphasis on risk-informed prioritisation aligns directly with AG-705's multi-factor scoring approach. Organisations mapping to NIST CSF can demonstrate PR.IP-12 compliance by showing that their AI agent's prioritisation incorporates the risk factors specified in Requirement 4.1 and that prioritisation decisions are documented per Requirement 4.3.

10. Failure Severity

FieldValue
Severity RatingCritical
Blast RadiusOrganisation-wide — incorrect prioritisation affects the residual risk posture across all systems and can lead to exploitation of the most exposed assets while resources are consumed on lower-risk remediation

Consequence chain: The agent misprioritises patch remediation due to reliance on a single severity metric, absence of dependency analysis, or failure to incorporate threat intelligence. The immediate effect is that high-risk vulnerabilities — those actively exploited, on internet-facing assets, or affecting safety-critical systems — remain unpatched while remediation resources are consumed by lower-risk items. This misallocation is invisible to standard compliance dashboards, which report "number of vulnerabilities patched" without distinguishing risk-weighted reduction. The organisation's reported patching metrics may show improvement while actual risk exposure increases. When an attacker exploits the deprioritised vulnerability, the blast radius is determined by the asset's criticality and network position — precisely the factors the agent failed to consider. In healthcare, this leads to clinical service disruption and patient safety risk. In financial services, this leads to trading halts, regulatory fines, and financial losses. In government, this leads to citizen data exposure and erosion of public trust. The downstream regulatory consequence is severe: the organisation cannot demonstrate that its vulnerability management was risk-proportionate, which is the standard set by NIS2, DORA, PCI DSS, and ISO 27001. The failure is compounded when dependency-blind patching causes production outages, because the organisation then suffers both the exploitation of unpatched vulnerabilities and the operational disruption of poorly managed patches. Remediation costs typically exceed the cost of governed prioritisation by two to three orders of magnitude — the £11.4 million healthcare breach in Scenario A could have been prevented by a prioritisation model that correctly incorporated the £0 marginal cost of ingesting KEV data.

Cross-references: AG-001 (Operational Boundary Enforcement) defines the operational boundaries within which the agent's remediation actions must remain. AG-004 (Action Rate Governance) constrains the rate at which the agent can execute patches to prevent overwhelming change management capacity. AG-007 (Governance Configuration Control) governs the prioritisation policy and scoring weights as configuration artefacts. AG-019 (Human Escalation & Override Triggers) defines the escalation mechanisms invoked by Requirement 4.4 for high-impact decisions. AG-022 (Behavioural Drift Detection) detects when the agent's prioritisation behaviour drifts from its governed baseline. AG-029 (Data Classification Enforcement) ensures that vulnerability and asset data consumed by the agent is appropriately classified. AG-055 (Audit Trail Immutability & Completeness) ensures that prioritisation records are tamper-proof and complete. AG-210 (Multi-Jurisdictional Regulatory Mapping) governs the alignment of remediation timelines with the strictest applicable regulatory requirement across jurisdictions. AG-699 (SOC Triage Integrity Governance) ensures that triage decisions feeding into patch prioritisation are themselves governed. AG-704 (Threat Intel Source Reliability Governance) ensures the reliability of the threat intelligence sources consumed by the prioritisation model. AG-706 (Autonomous Remediation Approval Governance) governs the approval workflow for agent-initiated remediation actions downstream of prioritisation.

Cite this protocol
AgentGoverning. (2026). AG-705: Patch Prioritisation Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-705