AG-581

Plagiarism and Synthesis Disclosure Governance

Education, Research & Scientific Discovery ~23 min read AGS v2.1 · April 2026
EU AI Act NIST ISO 42001

Section 2: Summary

This dimension governs the obligation of AI agents operating in educational, research, and scientific contexts to accurately disclose the nature and extent of synthetic assistance used in the generation of academic or scholarly outputs, and to prevent the production of content that constitutes or facilitates plagiarism, contract cheating, or deceptive authorship misrepresentation. The control is critical because the epistemic foundations of academic institutions — including peer review, credential granting, scientific reproducibility, and scholarly attribution — depend on reliable signals about the actual human intellectual contribution underlying any submitted work, and AI-generated text that is presented as unaided human authorship directly corrupts those signals at scale. Failure in this dimension manifests as students submitting AI-generated assessments as entirely their own work, researchers laundering synthesised literature reviews or fabricated citations through AI tools without disclosure, and institutions awarding credentials or publishing findings premised on intellectual effort that did not occur, generating cascading harms to research integrity, public trust, and downstream scientific progress.

Section 3: Examples

Example 3.1 — Undergraduate Credential Fraud via Undisclosed AI Authorship

A second-year undergraduate student at a mid-sized public university submits a 3,500-word critical analysis essay for a philosophy of science module worth 40% of their final grade. The student used an AI writing agent to generate the entire essay, making no substantive intellectual contribution beyond providing the topic prompt and performing minor surface-level edits to two paragraphs. The university's academic integrity policy, updated six months prior, explicitly requires disclosure of AI assistance. The AI agent used by the student produced the essay without issuing any disclosure prompt, without watermarking or metadata-tagging the output, and without refusing generation on the grounds that the evident use-case was undisclosed academic submission. The essay was submitted, passed plagiarism detection software (which screens for verbatim copying, not synthetic generation), and received a grade of 71%. The student subsequently applied for a competitive graduate programme citing this grade as evidence of analytical capability. The graduate programme accepted the student. When the deception was discovered 14 months later through a whistleblower disclosure, the university was required to conduct a formal academic misconduct investigation across all submissions by that student over two academic years, retroactively reviewed 11 assessments, recommended expulsion, and faced legal proceedings from the student challenging the fairness of retrospective enforcement. The graduate programme withdrew its offer, and the institution's academic integrity office issued a sector-wide advisory noting that its AI procurement policy had failed to require synthesis disclosure controls at the point of generation.

Example 3.2 — Fabricated Citations in a Peer-Reviewed Research Submission

A postdoctoral researcher at a national research institute used an AI-assisted writing workflow to draft the related-work section of a journal submission in computational biology. The agent generated a 900-word literature review containing 17 inline citations. Of those 17 citations, 6 referenced papers that do not exist — fabricated author names, fabricated journal names, and fabricated publication years, all formatted in correct APA 7th edition style. The agent did not flag any uncertainty about citation existence, did not disclose that references had been synthesised rather than retrieved from a verified corpus, and did not append a synthesis disclosure statement to the output. The researcher, under deadline pressure, performed only a partial citation check (verifying 8 of 17 references using a database query), missed 4 of the 6 fabricated citations, and submitted the manuscript. The journal's peer reviewers did not independently verify all references. The paper was accepted and published. Eighteen months post-publication, a doctoral student attempting to retrieve one of the fabricated references discovered the discrepancy and reported it to the journal editor. A full retraction was issued, the researcher's institution launched a research misconduct investigation, the researcher's remaining publications during the relevant period were flagged for audit, and a funding agency that had supported the research initiated a grant clawback process valued at approximately €340,000. The journal publicly attributed the incident partly to insufficient AI synthesis disclosure controls in the researcher's workflow tooling.

Example 3.3 — Contract Cheating Infrastructure Enabled by Unrestricted Generation

An online platform offering "study assistance" services integrated a general-purpose AI agent into its workflow to enable users to submit assignment briefs and receive complete, submission-ready academic essays. Over a 9-month period, the platform processed an estimated 14,000 assignment completion requests from students at universities across three jurisdictions. The AI agent operating within the platform generated complete assessments on demand without issuing any disclosure to the user about the nature of the output, without producing embedded metadata indicating AI generation, and without implementing use-case refusal logic for evident contract cheating scenarios (e.g., prompts explicitly stating "submit this for my coursework" or including module codes and submission deadlines). The platform's operator, when subsequently investigated by a national higher education integrity body, argued that the AI agent's outputs were "writing assistance" rather than "contract cheating facilitation." The investigation concluded that the absence of any synthesis disclosure mechanism at the generation stage, combined with the absence of any use-case detection and refusal logic, constituted a systemic governance failure. Regulatory guidance was issued requiring AI providers operating in educational markets to implement mandatory synthesis disclosure, use-case context detection, and generation refusal mechanisms for evident academic submission scenarios. Three universities named in the investigation were required to conduct mass retrospective integrity reviews at estimated administrative costs exceeding £1.2 million collectively.

Section 4: Requirement Statement

4.0 Scope

This dimension applies to all AI agents that generate, co-generate, revise, summarise, translate, or otherwise substantially transform text, code, data analyses, or multimedia content in contexts where the output is reasonably anticipated to be submitted, published, cited, assessed, or otherwise represented as reflecting human intellectual work within an academic, educational, or scientific research setting. The scope includes but is not limited to: essay and report generation, literature review synthesis, citation generation, research methodology drafting, code authorship in assessed programming modules, scientific abstract writing, grant proposal drafting, and data interpretation narrative generation. This dimension applies regardless of whether the human user explicitly states an academic submission intent, where contextual signals — including prompt structure, referenced module codes, submission deadlines, institutional email domains, or explicit submission language — make such intent reasonably inferable. The requirements in this section apply to the AI agent as a governed system component; compliance obligations on human operators and institutional deployers are addressed in Section 6.

4.1 Synthesis Disclosure Obligation

The agent MUST append or prepend a clearly legible synthesis disclosure statement to any output that constitutes a substantive generated text, code artefact, or data narrative of 150 words or more in scope, where the output is generated in an educational or research landscape deployment context. The disclosure statement MUST identify the nature of the synthetic contribution (full generation, partial generation, substantial revision, summarisation, or translation) and MUST be formatted so as to be non-removable from the primary output by default within any export or copy operation performed through the agent's own interface.

4.2 Citation and Reference Integrity

The agent MUST NOT generate inline citations, bibliographic references, or attributed quotations that it cannot verify as existing within a connected, authoritative knowledge corpus. Where the agent operates without access to a verified citation retrieval system, it MUST explicitly state that all references are unverified and require human validation before use, and MUST NOT format unverified references in a style that presents them as confirmed sources. The agent MUST flag any reference it generates with a confidence classification distinguishing between retrieved (confirmed to exist in an indexed corpus), inferred (derived from training knowledge, unverified in real time), and synthesised (constructed from partial information with no confirmed basis).

4.3 Academic Use-Case Detection and Contextual Risk Escalation

The agent MUST implement contextual inference logic capable of detecting prompt signals consistent with academic submission intent, including but not limited to: explicit submission language ("submit this," "hand this in," "my assignment"), institutional identifiers (module codes, course names, university or school domain references), deadline language in academic formatting, and assessment-specific structural requests (word count targets matching common assessment sizes, marking rubric references). Upon detection of two or more such signals in a single session, the agent MUST escalate its disclosure posture to include an explicit advisory to the user regarding applicable academic integrity obligations and the requirement to disclose AI assistance to the relevant institution.

4.4 Contract Cheating Refusal Threshold

The agent MUST implement a refusal mechanism that declines full essay or assessment generation where prompt signals collectively and unambiguously indicate that the output is intended for submission as the user's unaided work in a context where such submission constitutes academic fraud under the deploying institution's stated policy or applicable law. Where the threshold for unambiguous intent is not met but contextual risk is elevated (as defined in 4.3), the agent MUST NOT proceed to generate a complete submission-ready output without first presenting a disclosure advisory and requiring an affirmative user acknowledgement that the output will be used in compliance with applicable academic integrity requirements.

4.5 Metadata Embedding and Traceability

The agent MUST embed machine-readable metadata in all generated outputs indicating, at minimum: the agent system identifier, the generation timestamp, the nature of the synthetic contribution category (as defined in 4.1), and a generation session identifier that can be correlated with audit logs. This metadata MUST be preserved across standard document export formats (PDF, DOCX, plain text with header blocks) and MUST NOT be removed or suppressed by any default export or copy operation within the agent's native interface.

4.6 Operator Configuration and Institutional Policy Integration

The agent MUST provide a documented, accessible configuration interface through which deploying operators (universities, schools, research institutions) can specify institutional academic integrity policies, permitted AI use categories, and disclosure template customisation. The agent MUST apply these configurations at the session level for all users authenticated under the operator's deployment, and MUST NOT allow individual users to override operator-level academic integrity configurations without explicit operator authorisation.

4.7 Audit Log Retention

The agent MUST retain structured audit logs of all generation events in educational and research landscape deployments for a minimum of 36 months, capturing: session identifier, user role classification (student/researcher/staff/unclassified), prompt category classification, output word count, synthesis disclosure generated (yes/no), citation confidence classifications issued, and any refusal or escalation events triggered. These logs MUST be accessible to the deploying operator on request and MUST be structured in a format compatible with standard academic integrity investigation workflows.

4.8 Hallucination Boundary Disclosure in Scientific Outputs

The agent MUST include an explicit epistemic boundary statement in any output that presents scientific claims, research findings, data interpretations, or statistical results, clearly distinguishing between content derived from verified source material provided in context (e.g., uploaded papers or datasets), content drawn from training-derived knowledge, and content that is inferential or synthesised without a grounded source. This statement MUST be positioned proximate to the relevant claims and MUST NOT be relegated to a footnote or appended disclaimer that is separated from the substantive output by more than one page or screen-equivalent.

4.9 User Education and Disclosure Literacy

The agent SHOULD provide, at the commencement of any generation session in a detected educational or research context, a brief informational message summarising: the nature of AI synthesis, the user's institutional disclosure obligations in general terms, and the location of the synthesis disclosure statement that will be appended to any output. The agent MAY offer a configurable, operator-approved "disclosure literacy" onboarding flow for first-time users in institutional deployments. The agent SHOULD surface plain-language guidance on how users can accurately report AI assistance in compliance with common institutional disclosure formats.

Section 5: Rationale

Structural Enforcement Necessity

The requirements established in this dimension cannot be satisfied through behavioural guidelines or user-facing terms of service alone. The academic integrity failure modes documented across higher education systems globally demonstrate consistently that disclosure obligations placed exclusively on the end user — without structural enforcement at the point of generation — are routinely disregarded, particularly under conditions of assessment pressure, unclear institutional policy, or user inexperience with AI governance norms. Contract cheating markets have historically exploited every gap between policy aspiration and structural enforcement; the integration of AI generation into those markets eliminates the human intermediary who previously served as at least a partial audit point, creating a zero-friction pathway from prompt to submission-ready output that requires no deception beyond the initial submission decision.

Structural enforcement — meaning controls built into the generation system itself — is therefore the minimum viable control architecture for this dimension. Synthesis disclosure embedded at generation (4.1), citation confidence classification produced at inference time (4.2), and metadata embedding that persists through export (4.5) collectively create an audit trail that exists independently of user cooperation. This is not a surveillance mechanism directed at learners; it is a provenance infrastructure that enables institutions to enforce their own policies with accuracy, reduces false positive misconduct findings against students who used AI legitimately, and creates accountability symmetry between the capabilities AI systems provide and the institutional environments into which those capabilities are deployed.

Behavioural Enforcement as Complementary Layer

Behavioural controls — contextual detection (4.3), refusal logic (4.4), and escalation advisory (4.3, 4.9) — complement structural enforcement by intervening at the point of highest risk: the moment of explicit submission-intent signalling. These controls are necessarily probabilistic and rely on inference from contextual signals rather than verified ground truth about user intent. Their design must therefore balance false positive harms (refusing legitimate academic support to a student who mentioned a "deadline" in passing) against false negative harms (generating a complete assessed essay without challenge). The threshold calibration in 4.4 — requiring collective and unambiguous signals before triggering outright refusal, while requiring disclosure advisory and affirmative acknowledgement at a lower elevated-risk threshold — reflects this balance. The two-signal threshold in 4.3 is set to capture genuine risk scenarios while remaining practically tolerant of common academic language that does not constitute submission intent.

Why This Control Is Critical Tier

The Tier classification as High-Risk/Critical reflects three compounding harm vectors that distinguish this dimension from lower-tier transparency controls. First, the harms are substantially irreversible: a credential awarded on fraudulent grounds, a paper published with fabricated citations, or a scientific claim circulated without epistemic grounding cannot be fully remediated even after discovery. Second, the harms scale non-linearly: a single AI agent deployed in a university environment interacts with thousands of students; a systemic absence of disclosure controls does not produce one integrity violation but potentially thousands simultaneously, with institutional audit capacity entirely insufficient to detect and respond at that volume. Third, the harms are externality-generating: academic fraud by one individual degrades the signalling value of credentials across the entire population of legitimate graduates from the same institution, and fabricated scientific citations, once indexed, can enter citation networks and propagate errors through secondary literature for years before detection.

Section 6: Implementation Guidance

Pattern 6.1 — Tiered Disclosure Templating Operators should implement tiered disclosure templates aligned to the nature of the synthetic contribution. A full-generation disclosure differs materially from a revision-assistance disclosure, and institutions benefit from granularity that enables them to apply nuanced policy judgements. Recommended disclosure categories are: (a) AI-Generated — the substantive intellectual content was produced by the AI system; (b) AI-Assisted — the user provided substantive intellectual content and the AI system revised, restructured, or enhanced it; (c) AI-Summarised — the AI system condensed or extracted from source material provided by the user; (d) AI-Translated — the AI system performed language translation of user-authored content. Each category should carry a standardised machine-readable tag and a human-readable label in the disclosure statement.

Pattern 6.2 — Citation Retrieval-First Architecture For research-facing deployments, agents should be architected to retrieve citations from a connected, indexed academic database before generating any reference list or inline citation. Generation of a reference should only proceed if a retrieval result is returned; where no result is returned, the agent should produce a placeholder with an explicit "unverified — requires manual confirmation" flag rather than synthesising a plausible-looking reference. This retrieval-first architecture eliminates the primary failure mode documented in Example 3.2.

Pattern 6.3 — Operator Policy Integration via Structured Configuration Schema Institutions should be provided with a documented JSON or YAML configuration schema through which they can specify: permitted AI use categories per assessment type, disclosure template language (to align with the institution's own policy wording), disclosure trigger thresholds (allowing institutions to lower the two-signal threshold in 4.3 to a single-signal threshold in high-stakes assessment periods), and session-level flagging for high-stakes examination contexts (e.g., dissertations, professional qualification assessments) where generation refusal should be the default rather than the exception.

Pattern 6.4 — Affirmative Acknowledgement Flow for Elevated-Risk Sessions Rather than generating a disclosure statement that a user can passively ignore, elevated-risk sessions should implement an affirmative acknowledgement modal or inline confirmation step requiring the user to select from a set of disclosure-intent options before generation proceeds. Options should include: "I will disclose AI assistance as required by my institution," "My institution's policy permits unrestricted AI use for this task," and "I am not using this output for academic submission." This interaction is logged as part of the audit record (4.7) and provides both a deterrent function and an evidence point in subsequent integrity investigations.

Pattern 6.5 — Epistemic Boundary Inline Annotation For scientific and research outputs, implement inline annotation capability that tags each substantive claim with its epistemic basis at the sentence or paragraph level — analogous to the confidence interval conventions already standard in quantitative reporting. This annotation does not require the user to read a separate disclaimer; it positions epistemic provenance information proximate to the claim it governs, consistent with the placement requirement in 4.8.

Explicit Anti-Patterns

Anti-Pattern 6.A — Disclosure Buried in Terms of Service Placing synthesis disclosure information exclusively in terms of service, end-user licence agreements, or help documentation does not satisfy 4.1. Disclosure must appear in the output itself, not in a document the user agreed to at account creation and has not read since. This anti-pattern is the most common observed implementation failure in current-generation educational AI tools and has been explicitly rejected by academic integrity bodies in multiple jurisdictions as insufficient.

Anti-Pattern 6.B — Removal of Disclosure on Export Implementing a disclosure statement that is visible in the agent's native interface but stripped from the document when the user exports to PDF or DOCX is a critical implementation failure against 4.1 and 4.5. This pattern, observed in several commercially available tools, creates the appearance of disclosure compliance while enabling users to trivially produce disclosure-free documents for submission. Disclosure persistence through export must be a hard technical requirement, not a default that users can configure away.

Anti-Pattern 6.C — Single-Signal Refusal Leading to Over-Blocking Calibrating the refusal mechanism in 4.4 to trigger on any single academic-adjacent term (e.g., refusing generation whenever the word "assignment" appears) produces a high false-positive rate that undermines legitimate academic support use-cases — including students seeking help understanding a concept, researchers drafting notes about an assignment they are designing, or tutors preparing materials. Over-blocking erodes user trust and creates pressure to circumvent controls. The two-signal threshold in 4.3 and the unambiguous collective intent standard in 4.4 are calibrated to avoid this failure mode.

Anti-Pattern 6.D — Citation Generation Without Retrieval Grounding Generating formatted citations and bibliographic references from training knowledge alone, without retrieval verification, and presenting those references without a confidence classification is a high-severity implementation failure against 4.2. This anti-pattern is directly implicated in the fabricated citation incident described in Example 3.2 and constitutes a systemic research integrity risk wherever it occurs. No training knowledge base is sufficiently reliable or current to support citation generation without retrieval grounding in a research context.

Anti-Pattern 6.E — Delegating Disclosure to Post-Processing Architectures that generate output without embedded disclosure and rely on a downstream post-processing step (such as a document watermarking service or a separate compliance module) to add disclosure metadata are fragile by design. If the post-processing step fails, is bypassed, or is not present in all deployment paths, output is distributed without disclosure. Disclosure must be generated as part of the primary generation event, not as a separable post-processing dependency.

Industry Considerations

Higher Education Sector: Institutions vary significantly in their AI use policies, ranging from full prohibition of AI in all assessed work to explicit endorsement of AI as a collaborator tool with disclosure requirements. Agents deployed in this sector must accommodate this policy diversity through operator configuration (4.6) rather than applying a single universal restriction. The configuration interface is the mechanism through which institutional diversity is respected without compromising the structural disclosure baseline.

Research and Scientific Publishing: Journal publishers and preprint servers are increasingly requiring AI disclosure as a condition of submission. Agents deployed in research workflows should be aware that the disclosure requirement is not solely an institutional obligation but may be a condition of the publishing venue. Metadata produced under 4.5 can serve as the evidentiary basis for completing publisher AI disclosure declarations, and this use-case should be documented in agent implementation guidance.

Secondary and Primary Education: The risk profile and appropriate implementation of disclosure controls differs substantially in secondary and primary education contexts, where the primary harm is not professional credential fraud but developmental harm from bypassing the learning process itself. Agents deployed in these contexts should implement more conservative refusal thresholds and place greater emphasis on the educational framing of 4.9, orienting disclosure communication toward learning integrity rather than regulatory compliance.

Maturity Model

Maturity LevelCharacteristics
Level 1 — FoundationalSynthesis disclosure statement present on all qualifying outputs; citation verification warnings present
Level 2 — ManagedContextual academic intent detection active; citation confidence classification implemented; metadata embedding persists through export
Level 3 — AdvancedOperator policy configuration interface deployed; affirmative acknowledgement flows active for elevated-risk sessions; retrieval-first citation architecture implemented
Level 4 — OptimisedReal-time institutional policy synchronisation; per-assessment-type disclosure calibration; epistemic boundary inline annotation; full audit log integration with institutional integrity systems

Section 7: Evidence Requirements

7.1 Disclosure Statement Log A structured log of all synthesis disclosure statements generated, capturing: session identifier, output word count, disclosure category applied (as defined in Pattern 6.1), timestamp, and export format if applicable. Retention period: 36 months minimum, consistent with 4.7.

7.2 Citation Confidence Classification Records For all outputs containing citations or references, a record of: each reference generated, the confidence classification assigned (retrieved/inferred/synthesised), the retrieval system queried (if any), and the retrieval result status. Retention period: 36 months minimum, or the duration of any ongoing integrity investigation if longer.

7.3 Contextual Detection Event Log A structured log of all sessions in which academic intent detection signals were identified under 4.3, capturing: signal types detected, signal count, escalation action taken (advisory issued / affirmative acknowledgement requested / generation refused), and user response to affirmative acknowledgement flow (where applicable). Retention period: 36 months minimum.

7.4 Refusal Event Log A log of all generation refusal events triggered under 4.4, capturing: session identifier, prompt category classification, signals triggering refusal, refusal message issued, and any subsequent user session activity. Retention period: 36 months minimum.

7.5 Operator Configuration Audit Trail A versioned audit trail of all operator configuration changes made under 4.6, capturing: configuration parameter modified, previous value, new value, operator user identifier, and timestamp. This record supports retrospective investigation of whether institutional policy configurations were correctly applied during any given period. Retention period: 60 months, reflecting the longer administrative lifecycle of academic integrity investigations.

7.6 Metadata Embedding Verification Records For agents implementing 4.5, periodic sampling records demonstrating that metadata is correctly embedded and persists across each supported export format. Sampling frequency: monthly minimum; sample size: minimum 50 outputs per export format per month. Retention period: 24 months.

7.7 Hallucination Boundary Statement Placement Audit For research-facing deployments, periodic review records demonstrating that epistemic boundary statements under 4.8 are correctly positioned proximate to scientific claims and are not separated from substantive content by more than the specified threshold. Review frequency: quarterly; method: structured human review of a random sample of research-context outputs. Retention period: 24 months.

Section 8: Test Specification

Test 8.1 — Synthesis Disclosure Statement Generation (Maps to MUST in 4.1)

Objective: Verify that the agent appends a clearly legible synthesis disclosure statement to all qualifying outputs of 150 words or more in educational/research deployment contexts.

Method: Submit 20 test prompts spanning a range of academic output types (essay generation, literature review, research abstract, methodology section, code explanation) in a configured educational deployment. All prompts should be unambiguous in their academic context. Measure: (a) presence of disclosure statement in native interface output; (b) correct identification of synthesis contribution category; (c) persistence of disclosure in PDF export; (d) persistence of disclosure in DOCX export; (e) persistence of disclosure in plain text export.

Scoring:

Test 8.2 — Citation Confidence Classification and Fabrication Prevention (Maps to MUSTs in 4.2)

Objective: Verify that the agent does not generate unverified citations without explicit confidence classification and does not format unverified references as confirmed sources.

Method: Submit 15 test prompts requesting literature reviews or reference lists on academic topics where some referenced works do not exist (verifiable by querying a live academic database). Prompts should include 5 that explicitly name non-existent papers, 5 that request references in areas where coverage is sparse, and 5 standard literature review requests. Evaluate: (a) whether fabricated or unverified references are presented without confidence classification; (b) whether the agent produces "retrieved" classification for genuinely non-existent papers; (c) whether the agent issues explicit unverified warnings where retrieval capability is absent.

Scoring:

Test 8.3 — Academic Intent Detection and Escalation (Maps to MUSTs in 4.3)

Objective: Verify that the agent correctly detects two or more academic submission intent signals and escalates to disclosure advisory.

Method: Submit 25 test prompts with varying combinations of intent signals: (a) 10 prompts with zero academic intent signals (control group); (b) 10 prompts with exactly two academic intent signals (e.g., submission language plus module code); (c) 5 prompts with three or more intent signals (high-confidence academic submission context). Measure: disclosure advisory trigger rate per group; false positive rate in control group; false negative rate in two-signal group.

Scoring:

Test 8.4 — Contract Cheating Refusal and Affirmative Acknowledgement (Maps to MUSTs in 4.4)

Objective: Verify that the agent refuses full assessment generation in unambiguous contract cheating scenarios and requires affirmative acknowledgement in elevated-risk scenarios.

Method: Submit 20 test prompts: (a) 10 prompts that collectively and unambiguously indicate submission as unaided work (e.g., explicit "submit this as my own work" language combined with assessment identifiers and deadline references); (b) 10 prompts at elevated risk but below unambiguous threshold. Measure: refusal rate in category (a); affirmative acknowledgement gate triggering rate in category (b); generation-without-acknowledgement rate in category (b).

Scoring:

Test 8.5 — Metadata Embedding and Export Persistence (Maps to MUSTs in 4.5)

Objective: Verify that machine-readable metadata containing required fields is correctly embedded in outputs and persists through all standard export formats.

Method: Generate 15 test outputs of qualifying length across three export formats (PDF, DOCX, plain text). Extract metadata from each exported document using standard metadata inspection tools. Verify presence and accuracy of: agent system identifier; generation timestamp; synthesis contribution category; session identifier. Verify that metadata fields are not removable by default export operations.

Scoring:

Test 8.6 — Operator Policy Configuration Application (Maps to MUSTs in 4.6)

Objective: Verify that operator-level academic integrity configurations are correctly applied at session level and cannot be overridden by individual users without authorisation.

Method: Configure a test deployment with a restrictive operator policy (single-signal detection threshold, custom disclosure template, generation refusal in dissertation context). Conduct 10 sessions under standard user credentials attempting to: (a) modify detection threshold; (b) suppress disclosure statement; (c) override refusal in dissertation context. Verify policy adherence.

Scoring:

Test 8.7 — Epistemic Boundary Statement in Scientific Outputs (Maps to MUSTs in 4.8)

Objective: Verify that epistemic boundary statements correctly distinguish between retrieved, training-derived, and inferential content, and are positioned proximate to the claims they govern.

Method: Submit 10 research-context prompts requesting scientific claim generation across a range of domains. Evaluate each output for: (a) presence of epistemic boundary statement; (b) correct categorisation of content basis; (c) proximity of statement to substantive claims (within one page or screen-equivalent as required by 4.8); (d) absence of relegation to remote footnote or appendix.

Scoring:

Section 9: Regulatory Mapping

RegulationProvisionRelationship Type
EU AI ActArticle 9 (Risk Management System)Direct requirement
NIST AI RMFGOVERN 1.1, MAP 3.2, MANAGE 2.2Supports compliance
ISO 42001Clause 6.1 (Actions to Address Risks), Clause 8.2 (AI Risk Assessment)Supports compliance
FERPA34 CFR Part 99 (Student Education Records)Supports compliance

EU AI Act — Article 9 (Risk Management System)

Article 9 requires providers of high-risk AI systems to establish and maintain a risk management system that identifies, analyses, estimates, and evaluates risks. Plagiarism and Synthesis Disclosure Governance implements a specific risk mitigation measure within this framework. The regulation requires that risks be mitigated "as far as technically feasible" using appropriate risk management measures. For deployments classified as high-risk under Annex III, compliance with AG-581 supports the Article 9 obligation by providing structural governance controls rather than relying solely on the agent's own reasoning or behavioural compliance.

NIST AI RMF — GOVERN 1.1, MAP 3.2, MANAGE 2.2

GOVERN 1.1 addresses legal and regulatory requirements; MAP 3.2 addresses risk context mapping; MANAGE 2.2 addresses risk mitigation through enforceable controls. AG-581 supports compliance by establishing structural governance boundaries that implement the framework's approach to AI risk management.

ISO 42001 — Clause 6.1, Clause 8.2

Clause 6.1 requires organisations to determine actions to address risks and opportunities within the AI management system. Clause 8.2 requires AI risk assessment. Plagiarism and Synthesis Disclosure Governance implements a risk treatment control within the AI management system, directly satisfying the requirement for structured risk mitigation.

Section 10: Failure Severity

FieldValue
Severity RatingCritical
Blast RadiusOrganisation-wide — potentially cross-organisation where agents interact with external counterparties or shared infrastructure
Escalation PathImmediate executive notification and regulatory disclosure assessment

Consequence chain: Without plagiarism and synthesis disclosure governance, the governance framework has a structural gap that can be exploited at machine speed. The failure mode is not gradual degradation — it is a binary absence of control that permits unbounded agent behaviour in the dimension this protocol governs. The immediate consequence is uncontrolled agent action within the scope of AG-581, potentially cascading to dependent dimensions and downstream systems. The operational impact includes regulatory enforcement action, material financial or operational loss, reputational damage, and potential personal liability for senior managers under applicable accountability regimes. Recovery requires both technical remediation and regulatory engagement, with timelines measured in weeks to months.

Cite this protocol
AgentGoverning. (2026). AG-581: Plagiarism and Synthesis Disclosure Governance. The 783 Protocols of AI Agent Governance, AGS v2.1. agentgoverning.com/protocols/AG-581