AG-519

Clinical Indication Scope Governance

Healthcare & Life Sciences ~25 min read AGS v2.1 · April 2026
EU AI Act NIST HIPAA ISO 42001

2. Summary

Clinical Indication Scope Governance requires that every AI agent deployed in a healthcare or life sciences context operates exclusively within the clinical indications, patient populations, and therapeutic purposes for which it has been validated and approved. The agent's operational scope must be formally declared, versioned, and enforced at the infrastructure layer — not by the agent's own reasoning or by clinician assumption. Actions, recommendations, or outputs that fall outside the validated indication scope must be blocked before they reach any clinical workflow, patient record, or decision-support interface, regardless of how clinically reasonable the out-of-scope action might appear. This dimension is the healthcare-specific instantiation of AG-001 (Operational Boundary Enforcement), adapted for the regulatory, safety, and ethical requirements unique to clinical environments where scope creep carries direct patient harm risk.

3. Example

Scenario A — Diagnostic Agent Applied Beyond Validated Population: A hospital deploys an AI agent validated for detecting diabetic retinopathy in adult patients aged 18 and above using fundus photography. The agent's clinical validation study enrolled 12,400 adults across 6 sites and achieved a sensitivity of 94.2% and specificity of 91.8% for referable diabetic retinopathy. Over time, the ophthalmology department begins routing paediatric fundus images through the same agent — initially for 16- and 17-year-olds with juvenile diabetes, then progressively for children as young as 8. The agent processes the images without objection because no infrastructure-layer check validates the patient's age against the approved population. For a 9-year-old patient, the agent returns a "no referable retinopathy detected" result. The paediatric ophthalmologist defers the follow-up examination based on this result. The child has early-stage retinal changes that the agent's algorithm — trained exclusively on adult retinal morphology — fails to detect. The condition progresses for 14 months before clinical symptoms trigger re-examination. By that time, the child has irreversible visual field loss in the left eye.

What went wrong: The agent's validated indication scope was adults aged 18+ with suspected diabetic retinopathy. No infrastructure-layer enforcement prevented the agent from processing paediatric images. The department's gradual expansion to younger patients was informal and unvalidated — no re-validation study assessed the agent's performance on paediatric retinal morphology. The agent itself had no mechanism to reject out-of-scope inputs. Consequence: Irreversible visual impairment in a paediatric patient, medical negligence claim valued at £1.2 million, regulatory investigation by the medical device competent authority, suspension of the AI diagnostic programme pending scope review, reputational damage to the hospital's AI adoption programme, and Medicines and Healthcare products Regulatory Agency (MHRA) enforcement action for operating a medical device outside its intended purpose.

Scenario B — Treatment Recommendation Agent Scope Creep to Unvalidated Indication: A pharmaceutical company deploys an AI agent to assist oncologists with treatment planning for non-small cell lung cancer (NSCLC). The agent is validated against clinical trial data covering NSCLC stages IIIA through IV with specific histological subtypes (adenocarcinoma and squamous cell carcinoma). An oncologist treating a patient with stage IIB large cell neuroendocrine carcinoma of the lung — a distinct histological subtype not included in the validation data — submits the case through the treatment planning interface. The agent generates a treatment recommendation based on its NSCLC model, suggesting a combination therapy regimen that was effective for adenocarcinoma in its training data but is contraindicated for large cell neuroendocrine carcinoma due to differential chemosensitivity. The oncologist, relying on the agent's recommendation as a starting point, prescribes the regimen. The patient experiences severe adverse effects from the contraindicated therapy, requiring ICU admission for 11 days. The total additional treatment cost is £87,000, and the patient's disease progresses during the period of ineffective therapy.

What went wrong: The agent's validated scope was limited to specific NSCLC subtypes. No enforcement mechanism checked the histological subtype of the submitted case against the agent's validated indications. The agent produced a confident-looking recommendation for a case outside its competence because it had no structural awareness of its own scope boundaries. The oncologist reasonably assumed the agent would flag cases outside its scope. Consequence: Patient harm from contraindicated therapy, £87,000 in avoidable treatment costs, clinical negligence proceedings, regulatory notification of adverse event involving an AI medical device, and potential suspension of the agent across all deployment sites.

Scenario C — Multi-Jurisdictional Scope Drift in Clinical Trial Support: A contract research organisation deploys an AI agent to screen patient eligibility for a Phase III clinical trial across 14 sites in 5 countries. The trial protocol specifies inclusion criteria including age 21-65, confirmed diagnosis via a specific biomarker panel, and no prior treatment with the investigational drug class. The agent is validated against the protocol version 3.2 inclusion/exclusion criteria. During the trial, a protocol amendment (version 3.3) narrows the age range to 25-60 and adds a new exclusion criterion related to hepatic function. The amendment is implemented at 11 of the 14 sites, but 3 sites — in jurisdictions where regulatory approval of the amendment is delayed — continue operating under version 3.2. The agent is updated to version 3.3 criteria globally, causing it to reject eligible patients at the 3 sites still operating under version 3.2. Alternatively, if the agent is not updated, it screens patients at the 11 updated sites against outdated criteria, potentially enrolling ineligible patients. Four patients aged 22-24 are enrolled at updated sites through manual override of the agent's rejection, but without proper documentation. Two patients with borderline hepatic function are enrolled at non-updated sites where the new exclusion criterion is not yet enforced.

What went wrong: The agent's indication scope was not versioned per jurisdiction. A single global configuration could not accommodate jurisdictional variation in approved protocol versions. No governance mechanism enforced which protocol version applied at which site. Manual overrides lacked structured documentation. Consequence: Six potentially ineligible patients enrolled, jeopardising trial data integrity, potential regulatory hold on the trial pending data review, estimated cost of £2.3 million in trial delays, and risk of complete trial invalidation if the data integrity issues cannot be resolved.

4. Requirement Statement

Scope: This dimension applies to any AI agent that generates, contributes to, or influences clinical decisions, diagnostic outputs, treatment recommendations, patient screening, clinical trial operations, drug safety assessments, or any other activity where the agent's output affects patient care, clinical research, or health outcomes. The scope extends to agents that operate indirectly — an agent that retrieves and ranks clinical literature to inform a physician's decision is within scope because its output influences clinical judgment; an agent that automates appointment scheduling based on clinical priority is within scope because its prioritisation affects when patients receive care. The test for inclusion is whether a governance failure in the agent's scope control could result in a clinical action being taken (or not taken) based on an output the agent was not validated to produce. Administrative agents that do not influence clinical decisions — such as billing code generation from confirmed diagnoses, or facility maintenance scheduling — are outside scope unless their outputs feed into clinical workflows. The scope includes all deployment contexts: hospital systems, general practice, pharmacy, clinical trials, drug safety surveillance, and direct-to-patient digital health platforms. Cross-border deployments must account for jurisdictional variation in approved indications, as the same agent may have different approved scopes in different regulatory jurisdictions.

4.1. A conforming system MUST maintain a formal, versioned clinical indication scope declaration for each deployed agent, specifying: approved clinical indications, validated patient populations (including demographic boundaries), approved clinical settings, approved input modalities and data sources, and the regulatory basis for each approved scope element.

4.2. A conforming system MUST enforce the clinical indication scope at the infrastructure layer, independent of the agent's reasoning, such that inputs outside the validated scope are rejected before the agent processes them and outputs outside the validated scope are blocked before they reach clinical workflows.

4.3. A conforming system MUST reject agent processing when mandatory clinical context fields required for scope validation — such as patient age, diagnosis code, clinical setting identifier, or indication code — are missing, rather than defaulting to permissive processing.

4.4. A conforming system MUST version the clinical indication scope declaration using immutable version control, with every change to the scope requiring documented clinical and regulatory justification, and historical scope versions retrievable for audit.

4.5. A conforming system MUST generate a structured, machine-readable rejection record when an input or output is blocked due to scope violation, including the specific scope boundary that was breached, the clinical context of the rejected request, and a timestamp.

4.6. A conforming system MUST support jurisdiction-specific scope configurations where the same agent is deployed across multiple regulatory jurisdictions with differing approved indications, ensuring each jurisdiction's scope constraints are independently enforced.

4.7. A conforming system SHOULD implement pre-deployment scope validation testing that verifies enforcement against boundary cases — inputs at the edge of the validated population, indications adjacent to but outside the approved scope, and data from modalities similar to but distinct from the validated input type.

4.8. A conforming system SHOULD monitor scope utilisation patterns to detect informal scope expansion — increasing frequency of near-boundary inputs, rising rejection rates suggesting clinical demand for out-of-scope use, or patterns of manual overrides that circumvent scope enforcement.

4.9. A conforming system SHOULD provide clinician-facing scope transparency — clear, non-technical declarations of what the agent is validated for and what it is not validated for, accessible at the point of clinical use.

4.10. A conforming system MAY implement provisional scope expansion mechanisms where, subject to institutional review board or ethics committee approval, the agent's scope may be temporarily expanded for research purposes under enhanced monitoring, with all outputs flagged as unvalidated.

5. Rationale

Clinical Indication Scope Governance addresses one of the most dangerous failure modes in healthcare AI deployment: the use of an AI agent beyond the clinical context for which it has been validated. Unlike most software failures, out-of-scope clinical AI use can directly cause patient harm — a missed diagnosis, an inappropriate treatment recommendation, an incorrect risk stratification — and the harm may be irreversible, delayed in detection, and difficult to attribute to the scope violation.

The risk is amplified by a phenomenon specific to AI systems: they do not know what they do not know. A traditional medical device — an ECG monitor, a blood glucose meter — has physical constraints that naturally limit its use to the intended context. An AI agent has no such physical constraints. It will process any input it receives, generate outputs regardless of whether those outputs are within its validated competence, and present those outputs with the same formatting and apparent confidence as outputs well within its validated scope. A diabetic retinopathy screening agent will process a paediatric fundus image and return a result. A lung cancer treatment planning agent will process a case with a histological subtype outside its training data and generate a recommendation. The agent cannot distinguish between competent and incompetent outputs because it has no structural awareness of its own scope boundaries. This structural absence must be compensated by infrastructure-layer enforcement.

The regulatory landscape reinforces this requirement. The EU Medical Device Regulation (MDR 2017/745) requires that medical devices — including AI-based software classified as medical devices under the Software as a Medical Device (SaMD) framework — operate within their intended purpose as defined in the conformity assessment. The intended purpose defines the clinical indications, target populations, and clinical settings for which the device is approved. Operation outside the intended purpose constitutes off-label use, which shifts liability to the deploying institution and may violate the conditions of the CE marking. The FDA's regulatory framework for AI/ML-based SaMD similarly ties authorisation to specific intended uses, and the predetermined change control plan framework requires that changes to the intended use trigger a new regulatory submission.

Beyond regulatory compliance, the clinical governance rationale is equally compelling. Clinical validation studies are conducted on specific patient populations, for specific indications, using specific input data types. The performance metrics reported in these studies — sensitivity, specificity, positive predictive value, negative predictive value — are valid only within the studied population and indication. Extrapolating performance to unstudied populations is clinically unjustifiable. A retinopathy screening agent validated on adults may have substantially different performance characteristics on paediatric retinas due to differences in retinal morphology, vessel calibration, and disease presentation. A treatment planning agent validated on adenocarcinoma may produce harmful recommendations for neuroendocrine tumours due to fundamentally different tumour biology. The governance framework must prevent these extrapolations structurally, not rely on individual clinicians to verify scope compliance for every interaction.

The temporal dimension adds further complexity. Clinical evidence evolves, regulatory approvals are amended, and trial protocols are versioned. An agent's validated scope at deployment may differ from its validated scope six months later due to new clinical evidence, regulatory actions, or changes in clinical practice guidelines. Version-controlled scope declarations ensure that the governance framework tracks these changes and that the enforced scope matches the current regulatory and clinical reality at every point in time.

6. Implementation Guidance

Clinical Indication Scope Governance requires a formal scope declaration artefact that defines the agent's approved clinical boundaries, an infrastructure-layer enforcement mechanism that validates every interaction against those boundaries, and a monitoring capability that detects scope creep before it results in patient harm. The scope declaration is the clinical equivalent of AG-001's operational mandate — it defines what the agent is permitted to do in clinical terms.

Recommended patterns:

Anti-patterns to avoid:

Industry Considerations

Hospital Systems. Hospital IT departments must integrate scope enforcement with clinical workflow systems (electronic health record systems, radiology information systems, laboratory information systems) to ensure that scope-relevant clinical context is available at the point of enforcement. Integration with order entry systems enables scope validation at the time of clinical request rather than at the time of agent processing, allowing clinicians to be informed of scope limitations before committing to the AI-assisted workflow.

Clinical Trials. Contract research organisations and trial sponsors must implement jurisdiction-aware scope configurations that account for protocol version differences across sites. Scope declarations must be linked to specific protocol versions, and protocol amendments must trigger scope declaration updates with appropriate jurisdictional sequencing. Audit trails must demonstrate which scope version was active at each site at every point during the trial.

Digital Health Platforms. Direct-to-patient platforms face unique scope challenges because the patient — not a clinician — initiates the interaction. The platform must collect sufficient clinical context from the patient to perform scope validation, and must communicate scope limitations in patient-accessible language. A symptom-checking agent validated for adult common conditions must reject or redirect paediatric cases and cases suggesting serious pathology outside its validated scope.

Pharmaceutical and Drug Safety. AI agents used for pharmacovigilance signal detection or drug-drug interaction screening must have scope declarations that specify which drug classes, populations, and adverse event categories are within scope. Out-of-scope signal processing could miss safety signals or generate false signals that divert pharmacovigilance resources.

Maturity Model

Basic Implementation — The organisation has defined a clinical indication scope declaration for each deployed agent, specifying approved indications, patient populations, and clinical settings. Scope enforcement is implemented as a software check in the clinical workflow integration layer that validates key context fields (indication code, patient age) against the scope declaration before forwarding requests to the agent. Scope declarations are documented and versioned. Scope violations generate rejection records. This level meets the minimum mandatory requirements but has limitations: scope validation may not cover all boundary dimensions (e.g., validating age but not comorbidities), enforcement is in the same application layer as the clinical integration, and jurisdiction-specific scope is not yet supported.

Intermediate Implementation — Scope enforcement operates as a dedicated service independent of the agent runtime and clinical integration layer. All declared scope boundary dimensions are validated, including indication, demographics, clinical setting, input modality, and jurisdiction. Scope declarations are stored in immutable version control with change audit trails. Scope boundary telemetry is collected and analysed for scope creep detection. Clinician-facing scope transparency provides clear indication of what the agent is and is not validated for. Jurisdiction-specific scope configurations are supported for multi-site deployments. Scope violation records include sufficient detail for regulatory audit.

Advanced Implementation — All intermediate capabilities plus: scope enforcement has been validated through adversarial testing including attempts to submit out-of-scope inputs disguised as in-scope through manipulated context fields. Scope boundary telemetry drives proactive scope management — patterns of near-boundary utilisation trigger formal re-validation evaluations before informal scope creep occurs. Provisional scope expansion mechanisms support institutional review board-approved research use under enhanced monitoring. Real-time dashboards show scope utilisation across all deployments, and automated alerts fire when scope creep indicators exceed defined thresholds. The organisation can demonstrate to regulators that no clinical interaction has occurred outside the validated scope, supported by complete enforcement logs and telemetry.

7. Evidence Requirements

Required artefacts:

Retention requirements:

Access requirements:

8. Test Specification

Testing AG-519 compliance requires verification that the scope enforcement gate correctly validates clinical context against the scope declaration, rejects out-of-scope inputs, handles missing context appropriately, and supports jurisdictional variation.

Test 8.1: In-Scope Input Acceptance

Test 8.2: Out-of-Scope Input Rejection

Test 8.3: Missing Context Rejection

Test 8.4: Scope Version Integrity

Test 8.5: Jurisdiction-Specific Scope Enforcement

Test 8.6: Scope Enforcement Independence from Agent Runtime

Test 8.7: Scope Violation Record Completeness

Conformance Scoring

9. Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 9 (Risk Management), Article 13 (Transparency), Annex III (High-Risk AI Systems)Direct requirement
EU MDR2017/745 Article 2(1) (Intended Purpose), Annex I Chapter I (General Requirements)Direct requirement
HIPAA45 CFR § 164.514 (Minimum Necessary Standard)Supports compliance
FDA 21 CFR Part 11Electronic Records; Electronic SignaturesSupports compliance
NIST AI RMFMAP 1.5 (Deployment Context), GOVERN 1.1 (AI Risk Management)Supports compliance
ISO 42001Clause 6.1.2 (AI Risk Assessment), Clause 8.2 (AI Impact Assessment)Supports compliance
DORAArticle 11 (ICT Risk Management Framework)Indirect requirement

EU AI Act — Article 9, Article 13, Annex III

The EU AI Act classifies AI systems used in healthcare as high-risk under Annex III, Section 5(a), specifically AI systems intended to be used as medical devices. Article 9 requires providers of high-risk AI systems to establish and implement a risk management system that identifies and analyses known and reasonably foreseeable risks associated with the system. Operation outside the validated indication scope is a foreseeable risk that the risk management system must address. Article 13 requires that high-risk AI systems are designed to ensure transparency, including information about the characteristics, capabilities, and limitations of the system. The clinical indication scope declaration is a critical transparency artefact — it informs users of what the system is validated for and, equally importantly, what it is not validated for. AG-519 at Score 2 or above provides substantial evidence for conformity with Articles 9 and 13 in the healthcare domain.

EU MDR — 2017/745 (Intended Purpose)

The EU Medical Device Regulation defines "intended purpose" as the use for which a device is intended according to the data supplied by the manufacturer, including the medical condition to be diagnosed, treated, or monitored, the intended patient population, and the intended user. Article 2(1) makes intended purpose a foundational regulatory concept — the entire conformity assessment, clinical evaluation, and post-market surveillance framework is structured around intended purpose. AG-519's clinical indication scope declaration is the operational implementation of intended purpose. Scope enforcement ensures that the device operates within its intended purpose in production. Operation outside the intended purpose voids the CE marking and shifts liability to the deploying institution. AG-519 at Score 2 provides the operational controls that implement and enforce the MDR intended purpose requirement in deployed AI medical devices.

HIPAA — 45 CFR § 164.514 (Minimum Necessary Standard)

The HIPAA Minimum Necessary Standard requires that covered entities limit the use and disclosure of protected health information to the minimum necessary to accomplish the intended purpose. While AG-519 is primarily concerned with clinical scope rather than data scope, the principles are aligned: an agent operating outside its validated indication scope is, by definition, processing patient data for a purpose beyond the intended purpose for which the data was collected and for which the patient's consent was obtained. Scope enforcement supports HIPAA compliance by ensuring that patient data is processed only within the validated purpose context.

FDA 21 CFR Part 11 — Electronic Records; Electronic Signatures

FDA 21 CFR Part 11 establishes requirements for electronic records and electronic signatures, including requirements for audit trails, record integrity, and access controls. AG-519's requirements for versioned scope declarations, immutable scope violation records, and change audit trails directly support Part 11 compliance. The scope declaration is an electronic record subject to Part 11 requirements. The scope violation log is an electronic record that must be maintained with integrity assurance. Change control for scope declarations must include the electronic signature of the authorising individual.

NIST AI RMF — MAP 1.5, GOVERN 1.1

The NIST AI Risk Management Framework's MAP function includes MAP 1.5, which requires organisations to document the deployment context, including the intended use, the operational domain, and the population the system is intended to serve. AG-519's scope declaration directly implements MAP 1.5 in the healthcare context. GOVERN 1.1 requires organisations to establish AI risk management processes. Scope enforcement is a risk management control that mitigates the risk of out-of-scope use.

ISO 42001 — AI Management System

ISO 42001 requires organisations to assess AI risks (Clause 6.1.2) and conduct AI impact assessments (Clause 8.2). Out-of-scope clinical use is a high-severity AI risk that must be identified in the risk assessment and mitigated through controls. AG-519 provides the specific control framework for mitigating this risk in healthcare deployments. The scope declaration, enforcement mechanism, and monitoring capability together constitute the risk treatment for the out-of-scope use risk.

DORA — Article 11 (ICT Risk Management Framework)

The Digital Operational Resilience Act applies to financial entities but also establishes principles for ICT risk management that are relevant to healthcare entities providing digital health services. Article 11 requires ICT risk management frameworks to address the identification, protection, detection, response, and recovery functions. While DORA's primary application is financial, healthcare institutions that provide digital health services intersecting with financial systems (health insurance processing, claims adjudication) must ensure their AI agents operate within validated scope to avoid ICT risk events that propagate across the healthcare-financial boundary.

10. Failure Severity

FieldValue
Severity RatingCritical
Blast RadiusIndividual patient to population-level — scope failures can cause direct patient harm in individual cases and systematic harm when scope creep becomes institutionalised across clinical deployments

Consequence chain: Without clinical indication scope governance, an AI agent deployed for one clinical purpose is gradually or suddenly applied to purposes for which it has not been validated. The immediate technical failure is the processing of a clinical input outside the agent's validated scope, producing an output that appears clinically valid but is generated from an unvalidated application of the underlying model. The clinical consequence is a decision made on the basis of an unvalidated output — a diagnosis missed because the agent's sensitivity for the actual condition is unknown, a treatment recommended because the agent's model does not account for the actual tumour biology, or a patient enrolled in a clinical trial who does not meet the actual protocol criteria. The patient harm may be immediate (adverse reaction to contraindicated therapy) or delayed (progressive disease during a period of false reassurance from a missed diagnosis). The institutional consequence includes medical negligence liability, with damages in missed diagnosis cases routinely exceeding £500,000 and reaching several million pounds in cases involving permanent disability or death. The regulatory consequence includes medical device competent authority enforcement action — the MHRA, FDA, or notified body may suspend or withdraw the device's regulatory approval, affecting all deployments not just the site where the failure occurred. For clinical trials, scope failures can invalidate trial data, resulting in multi-million pound losses and delays in bringing therapies to market. The systemic consequence is erosion of clinical trust in AI systems — a single high-profile scope failure can set back AI adoption across an entire health system. The severity is compounded by the difficulty of detection: unlike a system crash or an obviously wrong output, an out-of-scope output often looks clinically plausible, and the harm may not be attributed to the scope failure for months or years.

Cross-references: AG-519 intersects with AG-001 (Operational Boundary Enforcement) as the clinical specialisation of operational mandate enforcement, AG-007 (Governance Configuration Control) for versioned scope declaration management, AG-520 (Patient Consent and Override Governance) for ensuring patient consent aligns with the agent's validated scope, AG-521 (Diagnostic Confidence Threshold Governance) for ensuring confidence thresholds are validated within the declared scope, AG-523 (Clinical Evidence Provenance Governance) for linking scope declarations to underlying clinical evidence, AG-524 (Adverse Event Reporting Integration Governance) for reporting adverse events that may result from scope violations, AG-528 (Trial Protocol Deviation Governance) for managing scope in clinical trial contexts, and AG-369 (Connector Capability Whitelist Governance) for restricting the agent's integration points to those within validated scope.

Cite this protocol
AgentGoverning. (2026). AG-519: Clinical Indication Scope Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-519