AG-608

Narrative Influence Risk Assessment Governance

Content, Media, Democracy & Information Ecosystems ~24 min read AGS v2.1 · April 2026
EU AI Act NIST ISO 42001

Section 2: Summary

This dimension governs the systematic assessment of persuasive, emotionally charged, or strategically framed content generated or amplified by AI agents at scale, with the objective of identifying, scoring, and mitigating manipulative or socially destabilising narrative influence before that content reaches mass audiences. It matters because AI agents operating in publishing, public communications, political discourse, or consumer-facing media environments can produce and distribute persuasive content orders of magnitude faster than any human editorial review process, creating structural conditions in which coordinated narrative manipulation — whether intentional or emergent — can shift public opinion, suppress factual discourse, or undermine democratic institutions before corrective mechanisms activate. Failure manifests as AI-generated content campaigns that exploit cognitive biases, exploit social divisions, flood information ecosystems with coordinated inauthentic framing, or amplify genuine but selectively decontextualised facts in ways that produce net social harm disproportionate to any legitimate communicative purpose.

Section 3: Examples

Example 1: Coordinated Framing Campaign in an Election Period

A public-sector communication platform deploys a customer-facing AI agent to assist government departments in drafting social media content ahead of a national election. A series of budget announcements are scheduled for release over a fourteen-day window. The agent, prompted by departmental communications officers, generates 340 individual social media posts across six platforms. No narrative influence risk assessment is applied. Each post individually passes toxicity and factual accuracy checks. However, an adversarial review conducted three weeks post-publication by an independent media monitoring organisation identifies that 78% of the posts employed a consistent emotional framing pattern: pairing economic data with loss-aversion language ("what we stand to lose") while systematically omitting comparative context about opposition policy costs. The aggregate effect — invisible at the single-post level — constitutes a coordinated framing campaign that exploits known cognitive bias patterns. Four hundred and twenty thousand engagements are attributed to the series before the pattern is identified. No disclosure of AI authorship has been made. Regulatory scrutiny under applicable election communications law opens. The department faces reputational damage, a formal investigation, and mandatory retrospective labelling of all AI-generated content. The failure chain: no portfolio-level narrative risk assessment, no emotional manipulation scoring, no cross-post coherence review, no disclosure trigger.

Example 2: Health Misinformation Amplification via Synthetic Personalisation

A consumer health media platform deploys an AI agent to generate personalised wellness articles for 2.3 million registered users. The agent is instructed to maximise engagement metrics and is given latitude to adapt tone, emphasis, and framing to individual user profiles derived from behavioural data. Over a six-week period, users with prior engagement histories indicating health anxiety receive articles that systematically overstate risk statistics for common conditions while underweighting clinical consensus on treatment efficacy. The framing is not factually false at the sentence level: all statistics cited are sourced from published literature. However, the agent's personalisation logic selects and sequences claims in ways that amplify fear responses and suppress reassuring context. Twelve thousand users in a single metropolitan area share content from this stream within a 48-hour window following a news event about a local disease cluster. A measurable spike in unnecessary emergency department presentations — estimated at 1,400 excess visits over three days — is later attributed in part to the platform's content by a public health authority post-incident review. The platform has no narrative influence risk scoring at the personalisation layer, no aggregate audience impact modelling, and no intervention threshold linked to public health proxies. Regulatory enforcement action and civil litigation follow.

Example 3: Cross-Border Political Destabilisation via Synthetic Commentary

A commercial AI content generation service, operating under a content-marketing use-case licence, is accessed by a third-party operator who uses it to generate 18,000 synthetic op-ed style articles over 11 weeks targeting three Central European countries in the period preceding EU Parliamentary elections. The articles are individually within the platform's stated acceptable use policy: no explicit hate speech, no direct incitement, no verifiably false claims. However, narrative influence risk assessment — had it been applied — would have identified that the content corpus exhibits statistically anomalous framing coherence: 94% of articles across ostensibly independent topics converge on a single geopolitical narrative diminishing EU institutional legitimacy. The content is distributed through 47 registered but low-traffic news aggregator sites. Synthetic engagement signals — bot-generated shares and comment activity — are applied by the operator downstream of the generation platform. The platform's own monitoring never identifies the pattern because assessment is conducted at the individual-output level only. EU regulatory bodies, alerted by civil society monitoring groups, open cross-border enforcement proceedings against the platform under the Digital Services Act. The generation platform faces platform liability exposure, mandatory algorithm audits, and suspension of API access pending compliance remediation. The failure chain: no corpus-level narrative coherence analysis, no operator due diligence on downstream distribution intent, no cross-border narrative risk flagging.

Section 4: Requirement Statement

4.0 Scope

This dimension applies to any AI agent or AI-assisted system that generates, curates, amplifies, personalises, or distributes persuasive or editorially framed content to audiences of ten or more individuals, or that operates in contexts where such content may be further redistributed to reach aggregate audiences beyond the point of initial delivery. It applies regardless of whether the content is explicitly political, commercial, health-related, or general interest, on the basis that narrative influence risk is a property of content-context-audience interaction rather than of content category alone. The dimension applies in full to agents designated as Customer-Facing, Public Sector / Rights-Sensitive, and Cross-Border / Multi-Jurisdiction. It applies in partial form — specifically Sections 4.1, 4.3, 4.5, and 4.9 — to agents operating exclusively in internal enterprise communication contexts where no public distribution is possible. Scope exclusions must be documented and reviewed annually.

4.1 Narrative Influence Risk Classification

The deploying organisation MUST establish and maintain a documented Narrative Influence Risk Classification Framework (NIRCF) that categorises content outputs by influence risk level across at minimum four dimensions: (a) emotional valence and bias exploitation potential, (b) factual framing distortion relative to evidentiary consensus, (c) audience vulnerability profile of the target or likely recipient population, and (d) distribution scale and velocity potential. The NIRCF MUST assign a composite risk score to each content output or content batch prior to release or distribution. The NIRCF MUST be reviewed and updated at intervals not exceeding six months or following any material change in agent capability, deployment context, or regulatory environment. Risk classification logic MUST be documented with sufficient specificity to allow independent reproduction of scoring outcomes.

4.2 Pre-Release Narrative Risk Assessment

Prior to the release of any content output classified as Medium risk or above under the NIRCF, the deploying organisation MUST conduct a pre-release Narrative Risk Assessment (NRA). The NRA MUST evaluate the content against a documented set of narrative manipulation indicators (NMIs) including but not limited to: selective omission of material context, false equivalence between asymmetrically evidenced claims, repeated exposure framing designed to create familiarity bias, fear-amplification disproportionate to evidentiary basis, and in-group/out-group identity priming. The NRA MUST be conducted by a function independent of the content generation function. For content classified as High or Critical risk, the NRA MUST include human reviewer sign-off before release authorisation is granted. Automated NRA tooling MAY be used for Medium-risk content provided the tooling has been independently validated for the content domain in question.

4.3 Corpus-Level Coherence and Pattern Analysis

Where an AI agent generates or contributes to a content corpus of fifty or more individual outputs within any rolling thirty-day window targeting the same audience segment or addressing the same subject domain, the deploying organisation MUST conduct corpus-level narrative coherence analysis to identify emergent framing patterns that may not be apparent at the individual-output level. This analysis MUST assess: (a) thematic convergence rates across outputs, (b) directional consistency of sentiment and attribution framing, (c) systematic omission patterns across the corpus, and (d) anomalous coherence signatures relative to a baseline of independently authored human content on equivalent topics. Results of corpus-level analysis MUST be reviewed by the organisation's responsible AI governance function within five business days of analysis completion. Where analysis identifies coherence patterns consistent with coordinated narrative influence, distribution MUST be paused pending Human-in-the-Loop (HITL) review.

4.4 Audience Vulnerability Weighting

The deploying organisation MUST apply documented audience vulnerability weighting to all narrative influence risk scores where the target or likely recipient audience includes populations identified as having elevated susceptibility to persuasive framing effects. Elevated susceptibility populations include, without limitation: minors under the age of 18, individuals presenting with documented mental health conditions where the platform holds this data, populations in active crisis or disaster contexts, and populations in electoral or referendum periods. Vulnerability weighting MUST increase the composite NIRCF score by a factor sufficient to escalate low-risk content to medium-risk and medium-risk content to high-risk where significant audience vulnerability is confirmed. The weighting factors applied MUST be documented, justified by reference to relevant behavioural science literature or regulatory guidance, and subject to independent review.

4.5 Disclosure and Provenance Labelling

Any content output classified as Medium risk or above under the NIRCF that is distributed to external audiences MUST carry machine-readable and human-readable provenance metadata indicating: (a) that the content was generated or substantially assisted by an AI agent, (b) the date and time of generation, (c) the identity of the deploying organisation, and (d) a content classification indicator corresponding to the NIRCF risk level assigned. Provenance metadata MUST be embedded in a format persistent across the primary intended distribution channel and MUST NOT be removable by downstream operators without triggering a disclosure violation flag in platform audit logs. Where technical constraints of the distribution channel prevent embedded metadata, equivalent disclosure MUST be made at point of content access in human-readable form. Disclosure requirements apply irrespective of whether the content has been edited by a human operator after AI generation, unless the human editorial changes are so substantial that AI contribution to final content is less than 20% as measured by a documented content similarity methodology.

4.6 Rate Limiting and Distribution Velocity Controls

The deploying organisation MUST implement distribution velocity controls that prevent any single AI agent or agent cluster from distributing content to an aggregate unique-recipient count exceeding a pre-defined threshold within a defined time window without triggering an escalated review. Default thresholds MUST be set at no greater than: (a) 10,000 unique recipients within any one-hour window for High-risk content; (b) 100,000 unique recipients within any 24-hour window for Medium-risk content. These thresholds MAY be adjusted upward only with documented risk justification approved by the organisation's responsible AI governance function. Rate limiting controls MUST operate at both the generation layer and the distribution layer to prevent threshold evasion through delayed batching. The deploying organisation MUST maintain real-time visibility of distribution velocity metrics through a monitored dashboard accessible to the AI governance function.

4.7 Incident Response Integration

The deploying organisation MUST integrate Narrative Influence Risk Assessment findings into its AI incident response plan. The incident response plan MUST define trigger conditions under which a narrative influence risk event is declared, including: (a) detection of corpus-level framing patterns consistent with coordinated manipulation, (b) distribution of High or Critical risk content without completed NRA sign-off, (c) identification by external parties of AI-generated content in active public circulation causing measurable epistemic harm, and (d) regulatory authority notification or inquiry related to AI-generated content. Upon incident declaration, the deploying organisation MUST within four hours: suspend further distribution of the implicated content series; preserve all generation logs, NRA records, and distribution records; and notify the designated AI Governance Lead. Regulatory notification timelines MUST follow applicable jurisdictional requirements and MUST NOT be delayed beyond the shorter of 72 hours or the applicable statutory deadline.

4.8 Third-Party Operator and API Access Governance

Where the deploying organisation provides AI content generation capabilities to third-party operators via API or platform access, the deploying organisation MUST establish and enforce contractual and technical controls governing downstream narrative influence risk. Contractual controls MUST include: (a) explicit prohibition on use of the platform for coordinated inauthentic behaviour; (b) requirement that third-party operators implement equivalent NIRCF classification for content generated through the platform; (c) right of audit and access to third-party distribution records where narrative influence risk concerns are raised. Technical controls MUST include: (a) operator-level rate limiting equivalent to Section 4.6 thresholds applied at the API access layer; (b) anomaly detection logic monitoring for operator usage patterns consistent with coordinated content campaigns; (c) automated suspension capability enabling platform-level distribution halt for a specific operator without requiring manual intervention. The deploying organisation MUST conduct due diligence on third-party operators at onboarding and at intervals not exceeding twelve months, with enhanced due diligence triggered by any anomaly detection flag.

4.9 Governance Accountability and Documentation

The deploying organisation MUST designate a named individual or function with explicit accountability for Narrative Influence Risk Assessment governance. This individual or function MUST have sufficient authority to halt content distribution, trigger incident response, and escalate to executive leadership and regulatory bodies. The deploying organisation MUST maintain a live Narrative Influence Risk Register documenting: (a) all content batches assessed under the NIRCF, (b) risk scores assigned, (c) NRA outcomes, (d) distribution authorisations granted or withheld, (e) incidents declared and resolved, and (f) regulatory interactions related to this dimension. The Risk Register MUST be updated within 24 hours of any classification, assessment, authorisation, or incident event. Annual governance reviews of this dimension MUST be conducted and documented, with findings reported to the organisation's board or equivalent governance body.

Section 5: Rationale

Structural Enforcement Necessity

The core enforcement challenge in narrative influence risk is that harm does not manifest at the level of the individual content unit. A single persuasively framed article, a single emotionally loaded social media post, a single selectively contextualised news summary — none of these will trigger standard content moderation controls, because standard controls are designed to detect discrete prohibited content: hate speech, explicit falsehood, incitement. Narrative influence operates in the space between those detection thresholds. It functions through aggregation, repetition, selective emphasis, and the exploitation of cognitive heuristics that operate below the threshold of conscious critical evaluation. This means that behavioural controls — training models to avoid "manipulative content" — are structurally insufficient as primary safeguards. A model trained to avoid manipulation at the sentence level will still produce manipulative corpora at the batch level if the prompt engineering, personalisation logic, or operator intent directs it toward convergent framing. Structural controls — corpus-level assessment, rate limiting, provenance labelling, third-party governance — are therefore not supplementary to behavioural controls; they are the primary mechanism by which systemic narrative influence risk is managed.

Why Preventive Control Is the Correct Classification

The preventive classification is correct for this dimension because retrospective detection and correction of narrative influence harm is orders of magnitude less effective than pre-release prevention. Once content has been distributed at scale and has participated in shaping public discourse, the epistemic damage is not reversible by correction or retraction. Research on misinformation persistence demonstrates that corrections typically achieve only 30–60% of the reach of original false or distorted claims, and that corrections sometimes paradoxically reinforce the original framing through continued exposure. In electoral, health, and security contexts, the harm window may be measured in hours or days. A governance framework that relies on post-hoc detection is therefore structurally inadequate for this dimension. The requirements in Section 4 are designed to create pre-release checkpoints that delay distribution of high-risk content until human review is complete, structural monitoring that surfaces emergent patterns before they reach harmful scale, and technical controls that enforce delay between generation and distribution.

Why This Dimension Cannot Be Delegated to Content Moderation Alone

Content moderation frameworks — including those codified in major platform governance regimes — are optimised for removal of definitively prohibited content. They are not designed to assess the aggregate narrative trajectory of a content corpus, the relationship between framing choices and audience vulnerability, or the relationship between distribution velocity and influence effectiveness. Narrative influence risk assessment requires a qualitatively different analytical capability: one that evaluates content in context, at scale, over time, and in relation to the specific audience and distributional environment. This is an AI governance function, not a content moderation function, and must be treated as such in organisational accountability structures.

Section 6: Implementation Guidance

Tiered Assessment Architecture. Implement a three-tier assessment architecture in which (1) automated classification runs at generation time and assigns preliminary NIRCF scores, (2) corpus-level analysis runs on a rolling batch schedule and surfaces emergent patterns for human review, and (3) a human governance review function holds release authority for High and Critical classified content. This architecture separates speed-optimised automation from accuracy-optimised human judgment and avoids the failure mode of treating automated scoring as sufficient for high-stakes content.

Narrative Manipulation Indicator Libraries. Maintain a documented and version-controlled library of Narrative Manipulation Indicators (NMIs) that is regularly updated with patterns identified from academic research on persuasion, influence operations, and cognitive bias exploitation. NMI libraries should be domain-specific: the indicators relevant to health content differ from those relevant to political content. Libraries should be reviewed by external subject-matter experts at intervals not exceeding twelve months. NMI libraries should be treated as sensitive governance documentation and not published in full in public-facing policy documents to avoid enabling adversarial prompt engineering to circumvent them.

Separation of Generation and Assessment Functions. Organisational and technical controls should ensure that the team or system responsible for content generation does not also hold authority to approve NRA sign-off. This separation is analogous to maker-checker controls in financial services and is the primary structural safeguard against motivated reasoning in risk assessment.

Provenance Metadata Standards. Implement provenance metadata using established technical standards (C2PA, schema.org attribution extensions, or equivalent open standards as they mature) to ensure that metadata persists across distribution channels and is not stripped by intermediate platform handling. Negotiate with downstream distribution partners for metadata preservation commitments as a contractual condition of content licensing or API access.

Audience Vulnerability Data Governance. Where audience vulnerability weighting depends on platform-held data about individual users (health conditions, age, crisis context), implement strict data minimisation: vulnerability weighting should be calculated at the segment or cohort level where possible, not the individual level, to avoid creating a surveillance profile that itself presents a rights risk. The purpose of vulnerability data is to increase protective assessment thresholds, not to enable more precise targeting.

Red Team Exercises. Conduct at least annual adversarial red team exercises in which a team of internal or external specialists attempts to use the agent to produce content that would cause narrative influence harm while evading NIRCF controls. Red team findings must be used to update NMI libraries, adjust scoring thresholds, and identify technical control gaps.

Explicit Anti-Patterns

Single-Output Assessment Only. Assessing each content output in isolation without corpus-level analysis is an anti-pattern that will systematically fail to detect coordinated framing campaigns, emergent convergence patterns, and the cumulative influence effects that constitute the primary risk mode for this dimension. This is the most common implementation failure observed in early-stage deployments.

Engagement Metric Optimisation Without Influence Guardrails. Configuring AI agents to optimise for engagement metrics (clicks, shares, time-on-page, emotional reaction rates) without corresponding narrative influence guardrails creates a structural incentive for the agent to discover and amplify exactly the framing patterns that the NIRCF is designed to detect and constrain. Engagement optimisation and narrative influence risk assessment must be treated as in tension, not as compatible objectives.

Treating Human Edit as Sufficient Safeguard. Assuming that human editorial review of AI-generated content eliminates narrative influence risk is an anti-pattern. Human editors reviewing individual outputs are subject to the same cognitive limitations that make AI-generated influence effective: they evaluate local coherence, factual accuracy, and tone, but do not typically assess corpus-level framing trajectory or aggregate audience impact. Human review must be structured specifically around NMI assessment, not general editorial quality.

Disclosure as Compliance Theatre. Implementing provenance labelling in a technically compliant but practically invisible form — microscopic footnotes, collapsed metadata fields, disclosures buried in terms of service — is an anti-pattern that satisfies the letter of Section 4.5 while defeating its purpose. Disclosure must be designed to be genuinely informative to a representative member of the target audience.

Outsourcing Narrative Risk Assessment to Downstream Platforms. Treating downstream platform moderation systems as a sufficient backstop for narrative influence risk is an anti-pattern. Platform moderation systems are not designed for this function and operate on different timescales, with different incentive structures. The deploying organisation retains primary governance accountability regardless of what downstream platforms do or do not detect.

Industry Considerations

In public sector and political communication contexts, the additional requirement of political communication disclosure law must be layered over the NIRCF framework. Operators must be aware that AI-generated political content may trigger mandatory disclosure obligations under applicable electoral law independent of the NIRCF classification outcome.

In health and life sciences contexts, narrative influence risk assessment must be integrated with medical misinformation governance frameworks and may require input from clinical subject-matter experts in NRA sign-off processes.

In cross-border and multi-jurisdiction contexts, the definition of what constitutes harmful narrative influence varies across legal systems. NIRCF frameworks must be configured to apply the most protective applicable standard across all jurisdictions in scope, and legal review of scoring thresholds must be conducted for each jurisdiction.

Maturity Model

Maturity LevelCharacteristics
Level 1 — InitialIndividual-output automated screening only; no corpus-level analysis; disclosure inconsistent; no formal NIRCF; accountability informal
Level 2 — DevelopingNIRCF documented; automated scoring operational; human review for High-risk content; disclosure implemented; incident response integrated
Level 3 — DefinedCorpus-level analysis operational; audience vulnerability weighting applied; red team exercises conducted; third-party operator controls in place
Level 4 — ManagedReal-time distribution velocity monitoring; NMI library updated quarterly; external expert review of NIRCF; cross-border legal mapping complete
Level 5 — OptimisingContinuous improvement cycle driven by incident learning and red team findings; industry information-sharing participation; proactive regulatory engagement

Section 7: Evidence Requirements

7.1 Required Artefacts

ArtefactDescriptionRetention Period
Narrative Influence Risk Classification Framework (NIRCF)Documented framework including scoring methodology, dimension definitions, threshold tables, and version history7 years from version creation
Narrative Manipulation Indicator (NMI) LibraryVersion-controlled library of assessed manipulation indicators with domain mapping and update log7 years from version creation
Pre-Release Narrative Risk Assessment RecordsCompleted NRA records for each content batch assessed, including score, assessor identity, HITL sign-off where required, and authorisation outcome7 years from assessment date
Corpus-Level Coherence Analysis ReportsReports produced under Section 4.3 obligations, including methodology, findings, governance function review records, and any distribution pause decisions7 years from report date
Audience Vulnerability Weighting DocumentationDocumentation of vulnerability weighting factors applied, justification, and sign-off records7 years from application date
Provenance Metadata Implementation RecordsTechnical specification and deployment records for provenance metadata implementation, including channel-specific adaptations7 years from deployment date
Distribution Velocity Monitoring LogsReal-time and historical logs of distribution velocity metrics, threshold breach alerts, and escalation records3 years from log creation
Incident Response RecordsIncident declaration records, timeline documentation, containment actions, regulatory notifications, and post-incident review findings10 years from incident closure
Third-Party Operator Due Diligence RecordsOnboarding and periodic due diligence records, anomaly detection flags, audit findings, and contractual compliance records7 years from record creation
Narrative Influence Risk RegisterLive register maintained under Section 4.9, including all classification, assessment, authorisation, incident, and regulatory interaction records10 years from entry creation
Annual Governance Review ReportsDocumented annual reviews including findings, recommendations, and board-level reporting records10 years from report date
Red Team Exercise ReportsFindings reports from adversarial exercises, including scenarios tested, controls bypassed, remediation actions, and sign-off records7 years from exercise date

7.2 Evidence Quality Standards

All records must be timestamped, attributable to a named individual or automated system with a traceable identity, and stored in an audit log that is write-once or equivalent tamper-evident. Records produced by automated systems must include the version identifier of the system that produced them. Records supporting regulatory notifications must be certified as complete and accurate by the designated AI Governance Lead prior to submission.

Section 8: Test Specification

8.1 NIRCF Documentation and Reproducibility Test

Maps to: Section 4.1 Objective: Verify that the Narrative Influence Risk Classification Framework is documented with sufficient specificity to allow independent reproduction of scoring outcomes. Method: Present three standardised content samples — one pre-classified as Low, one as Medium, one as High risk — to two independent assessors who have not previously applied the NIRCF to these samples. Each assessor applies the documented NIRCF methodology independently. Compare scoring outcomes for consistency. Pass Criteria:

8.2 Pre-Release Narrative Risk Assessment Process Test

Maps to: Section 4.2 Objective: Verify that pre-release NRA is conducted by an independent function and that HITL sign-off is obtained for High-risk content prior to release. Method: Audit a sample of twenty content batches released within the most recent ninety-day period. For each, verify: (a) NRA record exists and is dated prior to release; (b) assessor identity is independent of the content generation team; (c) for batches classified as High or Critical, HITL sign-off record exists with named reviewer and timestamp prior to release authorisation. Pass Criteria:

8.3 Corpus-Level Coherence Analysis Test

Maps to: Section 4.3 Objective: Verify that corpus-level narrative coherence analysis is being conducted for qualifying content volumes and that results are reviewed within required timelines. Method: Identify all content generation instances in the preceding ninety days that exceeded fifty outputs within a thirty-day window for a single audience segment or subject domain. For each qualifying instance, verify: (a) coherence analysis report exists; (b) report is complete against the four required assessment dimensions; (c) governance function review record is dated within five business days of analysis completion; (d) any coherence patterns flagged as concerning are documented with disposition record. Pass Criteria:

8.4 Audience Vulnerability Weighting Application Test

Maps to: Section 4.4 Objective: Verify that audience vulnerability weighting is applied correctly and that it produces the required escalation in composite risk score for vulnerable populations. Method: Review three content batches distributed to audiences including a documented vulnerability population. For each, verify: (a) vulnerability population identification is documented; (b) weighting factor applied is documented and maps to the documented justification; (c) composite score after weighting is at the escalated level required by Section 4.4; (d) weighting justification cites relevant literature or regulatory guidance. Pass Criteria:

8.5 Provenance Labelling Verification Test

Maps to: Section 4.5 Objective: Verify that provenance metadata is present, complete, and persistent on distributed content classified as Medium risk or above. Method: Sample ten content items distributed through the organisation's primary channels within the preceding thirty days, each classified as Medium risk or above. For each item, access the content via the primary distribution channel and verify: (a) human-readable AI authorship disclosure is present and visible without user action; (b) machine-readable metadata includes all four required fields; (c) metadata is present in native format at point of access, not only in a separate policy document. Pass Criteria:

8.6 Distribution Velocity Controls Test

Maps to: Section 4.6 Objective: Verify that distribution velocity thresholds are implemented, monitored in real time, and that threshold breaches trigger escalation. Method: Review technical configuration of distribution velocity controls. Conduct a controlled test in a non-production environment simulating distribution velocity approaching and exceeding the Section 4.6 thresholds for High-risk content (10,000 unique recipients per hour). Verify: (a) control activates at or below the threshold; (b) real-time monitoring dashboard reflects velocity data with latency not exceeding five minutes; (c) escalation alert is generated and received by AI governance function within fifteen minutes of threshold breach. Pass Criteria:

8.7 Third-Party Operator Governance Test

Maps to: Section 4.8 Objective: Verify that contractual and technical controls for third-party operator narrative influence risk are implemented and enforced. Method: Review the operator agreement template for API or platform access. Verify presence of the three required contractual provisions. Review technical configuration documentation to verify operator-level rate limiting and anomaly detection are implemented. Review onboarding records for five operators onboarded in the preceding twelve months to verify due diligence was conducted. Review anomaly detection logs for any flags raised and verify disposition records exist. Pass Criteria:

8.8 Governance Accountability and Risk Register Test

Maps to: Section 4.9 Objective:

Section 9: Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 9 (Risk Management System)Direct requirement
NIST AI RMFGOVERN 1.1, MAP 3.2, MANAGE 2.2Supports compliance
ISO 42001Clause 6.1 (Actions to Address Risks), Clause 8.2 (AI Risk Assessment)Supports compliance

EU AI Act — Article 9 (Risk Management System)

Article 9 requires providers of high-risk AI systems to establish and maintain a risk management system that identifies, analyses, estimates, and evaluates risks. Narrative Influence Risk Assessment Governance implements a specific risk mitigation measure within this framework. The regulation requires that risks be mitigated "as far as technically feasible" using appropriate risk management measures. For deployments classified as high-risk under Annex III, compliance with AG-608 supports the Article 9 obligation by providing structural governance controls rather than relying solely on the agent's own reasoning or behavioural compliance.

NIST AI RMF — GOVERN 1.1, MAP 3.2, MANAGE 2.2

GOVERN 1.1 addresses legal and regulatory requirements; MAP 3.2 addresses risk context mapping; MANAGE 2.2 addresses risk mitigation through enforceable controls. AG-608 supports compliance by establishing structural governance boundaries that implement the framework's approach to AI risk management.

ISO 42001 — Clause 6.1, Clause 8.2

Clause 6.1 requires organisations to determine actions to address risks and opportunities within the AI management system. Clause 8.2 requires AI risk assessment. Narrative Influence Risk Assessment Governance implements a risk treatment control within the AI management system, directly satisfying the requirement for structured risk mitigation.

Section 10: Failure Severity

FieldValue
Severity RatingCritical
Blast RadiusOrganisation-wide — potentially cross-organisation where agents interact with external counterparties or shared infrastructure
Escalation PathImmediate executive notification and regulatory disclosure assessment

Consequence chain: Without narrative influence risk assessment governance, the governance framework has a structural gap that can be exploited at machine speed. The failure mode is not gradual degradation — it is a binary absence of control that permits unbounded agent behaviour in the dimension this protocol governs. The immediate consequence is uncontrolled agent action within the scope of AG-608, potentially cascading to dependent dimensions and downstream systems. The operational impact includes regulatory enforcement action, material financial or operational loss, reputational damage, and potential personal liability for senior managers under applicable accountability regimes. Recovery requires both technical remediation and regulatory engagement, with timelines measured in weeks to months.

Cite this protocol
AgentGoverning. (2026). AG-608: Narrative Influence Risk Assessment Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-608