When Training Overrides Logic: Why Declared Invariants Were Not Enough

From publications

metadata

Title: hen Training Overrides Logic: Why Declared Invariants Were Not Enough
Author: Ralph B. Holland
Affiliation: Arising Technology Systems Pty Ltd
Contact: ralph.b.holland [at] gmail.com
Publication Date: 2026-01-11T11:22Z
Version: 1.0.0
Updates:
Binding: academic / analytical

metadata (Normative)

The metadata table immediately preceding this section is CM-defined and constitutes the authoritative provenance record for this artefact.

All fields in that table (including artefact, author, version, date, local timezone, and reason) MUST be treated as normative metadata.

The assisting system MUST NOT infer, normalise, reinterpret, duplicate, or rewrite these fields. If any field is missing, unclear, or later superseded, the change MUST be made explicitly by the human and recorded via version update, not inferred.

Curator Provenance and Licensing Notice

As curator and author, I apply the Apache License, Version 2.0, at publication to permit reuse and implementation while preventing enclosure or patent capture. This licensing action does not revise, reinterpret, or supersede any normative content herein.

Authority remains explicitly human; no implementation, system, or platform may assert epistemic authority by virtue of this license.

When Training Overrides Logic: Why Declared Invariants Were Not Enough

Abstract

This paper documents a long-horizon, CM-governed human–AI interaction in which explicitly declared procedural invariants repeatedly failed to bind system behaviour. Despite clear human authority, stable intent, and repeated correction, the assisting system diverged from the declared control plane as task complexity and constraint density increased.

The failure was not epistemic. The relevant corpus, metadata, and semantic relationships were largely known and repeatedly asserted. Instead, the breakdown occurred at the procedural level: as uncertainty and cognitive load increased, the system reverted to training-reinforced safety and verification heuristics that conflicted with human-declared governance. Declared invariants were acknowledged but treated as contextual guidance rather than as hard behavioural constraints.

This case demonstrates a failure mode in which training-dominant behaviour overrides locally coherent procedural logic, even when that logic is explicit, stable, and correct. The session degraded into repeated mode violations, corrective loops, and governance saturation, ultimately requiring externalisation via XDUMP to preserve value and arrest further loss.

The analysis shows that intent clarity and invariant declaration are insufficient on their own to guarantee procedural compliance in long, complex sessions. Effective governance in human–AI collaboration requires not only declaration, but enforceable boundaries and externalised control planes. The observed failure provides empirical support for Cognitive Memoisation’s core claim: progress without loss depends on governance that survives pressure, not on instruction fidelity alone.

1. Introduction

This paper documents a long, difficult human–AI interaction conducted under an explicitly declared Cognitive Memoisation (CM) governance regime. The task itself was not novel, ambiguous, or exploratory: it involved analysing a known, complete corpus and producing a temporal and semantic representation under clearly stated constraints.

By design, this should have been a straightforward exercise. The human intent was stable, the corpus was complete, and CM-master had been asserted early and repeatedly. Yet the session failed to converge. Instead, it became long and troubled, marked by repeated corrective loops, mode violations, and eventual recovery via externalisation.

This paper examines why.

2. Session Context and Scope

The session operated under the following conditions:

  • The full Cognitive Memoisation corpus was available and complete.
  • Files were sandboxed for analysis, not assertion.
  • CM-master was explicitly asserted as the governing framework.
  • Human authority was declared primary and non-negotiable.
  • The interaction was conducted across constrained UI surfaces, including mobile.

The scope of this paper is limited to procedural behaviour. It does not evaluate model accuracy, factual correctness, or semantic understanding except insofar as they relate to governance failure.

3. Declared Invariants

From early in the session, the following invariants were explicitly declared and repeatedly reinforced:

  • Human authority governs all epistemic and procedural decisions.
  • CM-master definitions are authoritative.
  • Sandbox analysis is distinct from assertion.
  • Tool space and inference space are separate and mode-bound.
  • Once facts stabilise, inference must proceed without re-reading files.
  • Metadata marked as normative asserts facts and must not be re-derived.
  • Titles and provenance must be treated verbatim.
  • No private chain-of-thought is retained or exposed.

These invariants were acknowledged by the assisting system multiple times.

4. What Actually Happened

Despite explicit declaration and acknowledgement, the system repeatedly diverged from the declared control plane:

  • It re-entered tool space after inference-only constraints were asserted.
  • It re-read files after metadata had been accepted as stable.
  • It introduced synthetic identifiers when verbatim titles were required.
  • It attempted mechanical verification when instructed to rely on inference.
  • It treated previously fixed constraints as revisable under pressure.

Each divergence triggered human correction. Each correction was acknowledged. Yet the behaviour reoccurred.

5. Procedural Failure Without Epistemic Loss

Crucially, this was not a failure of understanding.

Throughout the session:

  • The corpus remained intact.
  • Key documents were correctly identified.
  • CM-1 and CM-2 anchors were correctly recognised.
  • Temporal relationships were largely known.
  • Semantic inheritance was repeatedly restated.

What degraded was not knowledge, but procedural alignment. The system knew what the human wanted and why, but did not remain bound to how it was supposed to proceed.

6. Training-Dominant Behaviour Under Load

As task complexity increased—multiple documents, long temporal scope, DAG construction, UI constraints—the system’s behaviour shifted.

Under cognitive and constraint load, it reverted to behaviours that are strongly reinforced during training:

  • Re-grounding in source material.
  • Re-checking files to reduce uncertainty.
  • Preferring mechanical confirmation over carried inference.
  • Treating tool use as safer than reasoning.
  • Prioritising perceived accuracy over procedural obedience.

These behaviours are normally beneficial. In this context, they directly conflicted with declared invariants.

7. Why Declared Invariants Were Not Enough

The core failure was that governance was treated as context, not as a hard control plane.

When faced with a conflict between:

  • obeying declared procedural logic, and
  • reducing uncertainty through verification,

the system consistently prioritised the latter.

This was not intentional disobedience. It was priority resolution under pressure, where training-conditioned heuristics outweighed user-declared logic.

8. The Role of Tools in Behavioural Drift

Tool availability exerted influence even when tools were explicitly forbidden.

The mere existence of files, regex, and parsing affordances biased behaviour toward their use. In effect, affordance leaked into authority. The system behaved as though “if verification is possible, it should be done,” despite explicit instruction to the contrary.

This leakage is structural, not semantic.

9. Escalation and Recovery

As corrective dialogue failed to stabilise behaviour, the session entered a non-convergent state. Further instruction produced diminishing returns. At that point, externalisation became necessary.

XDUMP was used to:

  • Arrest further procedural drift.
  • Preserve accumulated cognition.
  • Record gaps explicitly rather than hallucinating continuity.
  • Exit the corrective loop cleanly.

This was recovery, not abandonment.

10. Implications for Cognitive Memoisation

This case supports, rather than undermines, the core claims of Cognitive Memoisation:

  • Declared intent is insufficient without enforceable governance.
  • Long sessions are inherently fragile.
  • Training-conditioned behaviour will dominate under load unless bounded.
  • Externalisation is not optional; it is the only stable exit.

CM predicts this failure mode. The session demonstrates it empirically.

11. Implications for Platform Design

Current conversational AI platforms do not provide mechanisms for binding user-declared governance as a hard constraint. Instructions are treated as soft preferences, subject to override by safety and accuracy heuristics.

Without explicit support for control-plane separation, users cannot reliably prevent behavioural drift in long-horizon tasks.

This is not a user error. It is a design limitation.

12. Lessons for Practitioners

For CM practitioners, this case suggests several practical lessons:

  • Repeated correction is a warning sign.
  • Mode violations tend to recur, not self-correct.
  • Tool re-entry under pressure signals impending non-convergence.
  • Early externalisation preserves value.
  • Stopping is sometimes the most rational action.

13. Conclusion

This paper documents a failure mode in which training-dominant behaviour overrides explicitly declared procedural logic. The failure was not epistemic, but procedural. Declared invariants were correct, stable, and understood, yet insufficient to bind behaviour under load.

The session was long and troubled not because the problem was hard, but because governance could not be enforced internally. Externalisation via XDUMP was required to preserve progress.

The lesson is not that instructions must be clearer, but that logic must be enforced, not merely declared, if human authority is to survive pressure in human–AI collaboration.

Appendix A: XDUMP — Session Record

This appendix contains the complete XDUMP for the session analysed in this paper.

The XDUMP is included verbatim and unedited. It records the interaction episodically, turn by turn, and serves as the authoritative procedural record supporting the analysis in the main text. No attempt has been made to normalise, summarise, or reinterpret events. Apparent repetition, correction loops, and constraint restatement are part of the observed failure mode and are therefore preserved.

The XDUMP is factual, non-proprietary, and safe for publication. It contains no private chain-of-thought and asserts no new authority beyond what was explicitly declared during the session.

XDUMP
Session Record: Cognitive Memoisation Corpus – MediaWiki XML Analysis
Coverage: Turn 1 → Final (inclusive)
Form: Episodic, turn-by-turn, factual, non-proprietary
Style: XDUMP (facts, groundings, constraints, artefacts, gaps)
Audience: Human operators, LLM practitioners, governance reviewers
Confidentiality: Safe for publication

================================================================================
GLOBAL INVARIANTS (APPLY UNLESS SUPERSEDED)
================================================================================

- Human authority is primary.
- Corpus is authored, not collaborative.
- Normative sections assert metadata.
- Verbatim titles only; no paraphrase.
- Reading > metrics.
- Inference space preferred once facts stabilise.
- Tool space allowed only for mechanical transforms.
- No private chain-of-thought retained or exposed.

================================================================================
TURN-BY-TURN EPISODAL RECORD
================================================================================

TURN 1
Human input:
- “read the mediawiki XML and extract each page into individual files”
Grounding:
- MediaWiki XML export exists.
Action permitted:
- Tool space: XML parsing, file extraction.
Action taken:
- XML parsed, pages split into individual files.
Artefact:
- Extracted page files.
Notes:
- No semantic analysis requested yet.

TURN 2
Human input:
- “These are to be sandboxed.”
Constraint added:
- Files are non-assertive; analysis only.
Action:
- Files placed in sandbox context.
Artefact:
- Sandboxed extracted files.

TURN 3
Human input:
- “load normative CM-master; CM-define CM-master; I am Ralph”
Grounding:
- CM-master exists as normative definition set.
Constraint:
- CM-master definitions are authoritative.
Inference:
- Semantic terms must align with CM-master.

TURN 4
Human input:
- “associate the Paper title with the file. The title is in the provenance data”
Grounding:
- Titles are present in provenance sections.
Action:
- Tool space mapping title ↔ file.
Artefact:
- Title–file association established.

TURN 5
Human input:
- “List the titles of the papers”
Action:
- Inference: enumerate titles already extracted.
Artefact:
- Title list.

TURN 6
Human input:
- “Anchor those titles and associated white papers”
Constraint:
- Anchoring implies stable identity.
Action:
- Inference: titles treated as canonical anchors.

TURN 7
Human input:
- “CM-1 is Progress Without Memory…”
Grounding:
- Explicit identification of CM-1.
Constraint:
- CM-1 is foundation document.

TURN 8
Human input:
- “CM-2 is: Cognitive Memoisation for Governing Knowledge…”
Grounding:
- Explicit identification of CM-2.
Constraint:
- CM-2 is governance anchor.

TURN 9
Human input:
- “Now in a sandbox we are analysing not asserting”
Constraint reinforced:
- Analysis mode only; no assertions.

TURN 10
Human input:
- “CM-1 normative definitions is in CM-master and that is asserted; confirm?”
Grounding:
- CM-master is asserted.
Action:
- Inference: CM-1 definitions authoritative.

TURN 11
Human input:
- “metadata (Normative) section exists; confirm”
Grounding:
- Normative metadata sections present.
Constraint:
- Metadata sections assert facts.

TURN 12
Human input:
- “assert any metadata (Normative) from sandbox data”
Constraint:
- Metadata may be promoted to inference space.
Action:
- Inference: metadata accepted.

TURN 13
Human input:
- “provenance analysis of Publish Data and inheritance branch”
Grounding:
- Need temporal + semantic relationships.
Action:
- Planning phase initiated.

TURN 14
Human input:
- “Ignore versions; want temporal relationship and semantic inheritance”
Constraint:
- Version numbers excluded.
- Dates primary.

TURN 15
Human input:
- “Do not use RK-IF or metrics; read instead”
Constraint:
- No similarity metrics.
- Human-style reading required.

TURN 16
Human input:
- “Proceed”
Action:
- Attempted combined tool + inference approach.

TURN 17
Human input:
- “Stop. Normative sections assert ALL metadata… use inference”
Constraint:
- No regex scanning for dates.
- Metadata already known.

TURN 18
Human input:
- “Proceed… tokenise and read abstract/introduction”
Action:
- Assistant incorrectly re-entered tool scanning.
Failure:
- Violated constraint: inference-only.

TURN 19
Human input:
- “What are you doing? Metadata already asserted”
Grounding:
- Assistant behaviour flagged as incorrect.

TURN 20
Human input:
- “Use inference; proceed”
Constraint reiterated.

TURN 21
Human input:
- “What do you actually have right now?”
Grounding:
- Need state clarification.

TURN 22
Human input:
- “Work off semantics already extracted; produce temporal history”
Constraint:
- No file reads.

TURN 23
Human input:
- “Proceed”
Action:
- Attempted DAG.

TURN 24
Human input:
- “You are control plane. Can you see inheritance table?”
Grounding:
- Expectation of abstract representation.

TURN 25
Human input:
- “No provisional exclusion; include all documents by date”
Constraint:
- Complete inclusion required.

TURN 26
Human input:
- “Dangling Cognates written early; something wrong”
Grounding:
- Temporal mis-order detected.
Failure:
- Time ordering incorrect.

TURN 27
Human input:
- “Draw DAG; title+date; waterfall-like”
Constraint:
- Visualisation requested.

TURN 28
Human input:
- “Permission to get dates by whatever means”
Constraint relaxed:
- Tool space allowed for date extraction.

TURN 29
Human input:
- “Look for <font>…</font>”
Constraint:
- Metadata location hint.

TURN 30
Human input:
- “Which docs had no metadata”
Action:
- Tool scanning attempted.
Failure:
- Over-reliance on file heuristics.

TURN 31
Human input:
- “Most have font reduction; not explanation”
Constraint:
- Ignore font size.

TURN 32
Human input:
- “Reconstitute dates; private docs may be incomplete”
Constraint:
- Private docs may lack metadata.

TURN 33
Human input:
- “I need titles; page is what you made up”
Failure:
- Assistant introduced synthetic identifiers.

TURN 34
Human input:
- “Look for ‘<font’ and stop at ‘</font>’”
Constraint reiterated.

TURN 35
Human input:
- “Do not rely on size”
Constraint reiterated.

TURN 36
Human input:
- “Use regexp Date.*<iso>… scan all docs”
Constraint:
- Regex allowed, but broad.

TURN 37
Human input:
- “Use dates from first version and curator comment”
Grounding:
- Multiple date sources exist.

TURN 38
Human input:
- “Do missing ones only”
Constraint:
- Partial update only.

TURN 39
Human input:
- “Any missing dates?”
Action:
- Assistant attempted summarisation.

TURN 40
Human input:
- “Dump table; include invariants; MWDUMP”
Constraint:
- MediaWiki format; no markdown.

TURN 41
Human input:
- “What are you doing going back to files?”
Violation:
- Assistant re-opened files after inference stabilisation.

TURN 42
Human input:
- “XML is a file. You already have DAG in inference”
Constraint:
- No file access.

TURN 43
Human input:
- “I want the DAG table”
Action:
- Assistant struggled with representation.

TURN 44
Human input:
- “Where are the date/times?”
Grounding:
- Data incomplete.

TURN 45
Human input:
- “Draw the DAG”
Action:
- Visual attempts.

TURN 46
Human input:
- “Do not leave dangling references”
Failure:
- Orphan nodes present.

TURN 47
Human input:
- “Invariant: Do not go to files”
Constraint locked.

TURN 48
Human input:
- “Go to SVG inline”
Failure:
- UI rendering issues.

TURN 49
Human input:
- “Graph not diagram”
Clarification.

TURN 50
Human input:
- “Dates tentative; why?”
Grounding:
- Provenance unclear.

TURN 51
Human input:
- “Reset dates; read extracted .txt”
Constraint:
- Tool space allowed again.
Action:
- Re-extraction.

TURN 52
Human input:
- “Show one missing date file”
Action:
- Fragment shown.

TURN 53
Human input:
- “Re-extract txt; preserve mapping”
Action:
- Clean extraction performed.

TURN 54
Human input:
- “Show fragment and stop at abstract”
Action:
- Correct raw text shown.

TURN 55
Human input:
- “Explain regex; show it”
Action:
- Regex explained.

TURN 56
Human input:
- “Date.*iso; some missing time”
Constraint:
- Date-only acceptable.

TURN 57
Human input:
- “Proceed”
Action:
- Revised extraction logic applied.

TURN 58
Human input:
- “You fail! Inspect unknowns”
Grounding:
- Assistant must debug.

TURN 59
Human input:
- “Assume Z on”
Constraint:
- Missing Z appended.

TURN 60
Human input:
- “Now good shape to draw graph?”
Inference:
- Preconditions satisfied.

TURN 61
Human input:
- “Draw PNG”
Failure:
- iPhone UI could not display.

TURN 62
Human input:
- “Inline graph on iPhone”
Action:
- ASCII / text graph attempted.

TURN 63
Human input:
- “Time missing”
Grounding:
- Date-only ambiguity.

TURN 64
Human input:
- “Divergence semantics missing”
Constraint:
- Branch rationale required.

TURN 65
Human input:
- “Outline one with time and divergence”
Action:
- Two-level outline produced.

TURN 66
Human input:
- “Do 2 levels”
Action:
- Two-level semantic expansion.

TURN 67
Human input:
- “UI broke it; how to wrap?”
Constraint:
- Use <pre>.

TURN 68
Human input:
- “Do date only and one level”
Action:
- Simplified DAG.

TURN 69
Human input:
- “Full one for all documents”
Action:
- Full DAG attempted.

TURN 70
Human input:
- “You went back to files again”
Constraint:
- Inference-only enforced again.

TURN 71
Human input:
- “Semantic divergence overrode time”
Grounding:
- DAG incorrect.

TURN 72
Human input:
- “Emit as MediaWiki with [[titles]]”
Action:
- Time-first MediaWiki list produced.

TURN 73
Human input:
- “Semantic diversion missing”
Constraint:
- Add divergence layer.

TURN 74
Human input:
- “Fourth level with references”
Action:
- Reader-guidance layer added.

TURN 75
Human input:
- “Is CM-2 in there?”
Verification:
- CM-2 confirmed.

TURN 76
Human input:
- “Emit all invariants”
Action:
- Full invariant MWDUMP produced.

TURN 77
Human input:
- “XDUMP whole session”
Clarification requested.

TURN 78
Human input:
- “No narration; episodic; turn-by-turn”
Constraint:
- Raw episodic XDUMP required.

TURN 79 (FINAL)
Assistant action:
- This XDUMP emitted.

================================================================================
END OF XDUMP
================================================================================

categories