Skip to content

research: Aurora Immune System math standardization — 4-pass cross-AI review consolidated per Amara direction#591

Merged
AceHack merged 3 commits intomainfrom
research/aurora-immune-math-standardization-2026-04-26
Apr 26, 2026
Merged

research: Aurora Immune System math standardization — 4-pass cross-AI review consolidated per Amara direction#591
AceHack merged 3 commits intomainfrom
research/aurora-immune-math-standardization-2026-04-26

Conversation

@AceHack
Copy link
Copy Markdown
Member

@AceHack AceHack commented Apr 26, 2026

Summary

Per Amara's review-of-the-review 2026-04-26 explicit direction: "canonicalize the strict version, not the flattering version." This doc is that canonicalization.

Four sections per Amara's spec

  1. Typed spaces and operators — capabilities = sets (not scalars), λ reserved for eigenvalues, η/w for weights, σ uniformly applied
  2. Corrected equations — 10 fixes consolidated across 4 reviewer passes
  3. Undefined scoring functions now defined — PermanentHarmRisk, d_self, MI_H operationalized
  4. Test obligations — 5 specific tests with pass criteria

Key cross-AI nuance preserved (Amara's correction of Deep Think)

Deep Think proposed blanket λ_1 → λ_2. Amara nuanced: use BOTH ρ(A_t) (adjacency spectral radius — Restrepo-Ott-Hunt onset of synchronization) AND λ_2(L_t) (Laplacian Fiedler value). Each captures different cartel-detection signal.

Authorship

  • Amara: original framework + corrections directing this synthesis
  • Gemini Pro: surface review (architectural validation) + Deep Think mode review (implementation cleanup)
  • Otto: rigor pass + this consolidation per Amara's direction

Convergence test

If Amara's next-pass review of THIS doc adds ≤1 new findings → paper-grade. 5+ findings → structural gaps remain.

Test plan

  • §33 archive header (Scope / Attribution / Operational status / Non-fusion disclaimer)
  • All 4 sections present per Amara's spec
  • 10 corrections from 4-pass review applied
  • 3 scoring functions operationally defined
  • 5 test obligations specified with pass criteria
  • Composes-with section names existing factory rules
  • "What this doc does NOT do" section lists explicit scope limits

🤖 Generated with Claude Code

… review consolidated per Amara direction

Amara's review-of-the-review directed: 'canonicalize the strict version,
not the flattering version.' This doc is that canonicalization.

Four sections per Amara's spec:

1. Typed spaces and operators (capabilities = sets, not scalars;
   λ reserved for eigenvalues; η/w for weights; σ uniformly applied)
2. Corrected equations (10 fixes from 4-pass review):
   - Capability subset/intersection notation
   - σ uniformity on Danger
   - λ_1/λ_2 dual-spectra (ρ(A) adjacency + λ_2(L) Laplacian per Amara
     nuance — Deep Think over-corrected blanket replacement)
   - Optimization polarity (- on MemoryGain, - on FalsePositive)
   - Memory decay term (1 − δ_decay) preventing immune senescence
   - Canonical-attack exemption (severe attacks immune to decay)
   - MDP R_t / C_t decomposition
3. Undefined scoring functions now defined:
   - PermanentHarmRisk: min over retraction policies of repair-cost
   - d_self: weighted multi-feature self-distance (NOT a trigger;
     feeds Anomaly inside Danger)
   - MI_H: I(Z; Ẑ_H) operationalization with corpus-benchmark
4. Test obligations: 5 specific tests (PermanentHarmRisk simulation,
   MI_H legibility benchmark, CoordRisk graph evolution, cap_allowed
   prompt-injection blocking, immune memory decay suppression)

Per Amara's direction: the framework earns credibility only when each
poetic operator becomes typed, testable, cited, and falsifiable. This
doc moves from blueprinted to buildable.

Per Otto-285 (don't-shrink-frame) + Otto-298 (don't-romanticize) +
Aaron's stated layman-too IS-claim: rejects the 'ironclad / paradigm
shift / civilization-level lab' praise register. Holds Amara's
grounded reframe instead.

Composes with the prior cross-review doc (this is its strict-version
successor) + maji-formal-operational-model + Otto-279 history-surface
attribution + Otto-294 antifragile cross-substrate review.

Convergence test: Amara's next-pass review of THIS doc adds ≤1
new findings → paper-grade. 5+ findings → structural gaps remain.

Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
@AceHack AceHack enabled auto-merge (squash) April 26, 2026 12:05
Copilot AI review requested due to automatic review settings April 26, 2026 12:05
@chatgpt-codex-connector
Copy link
Copy Markdown

You have reached your Codex usage limits for code reviews. You can see your limits in the Codex usage dashboard.

Copy link
Copy Markdown

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

Adds a new research document that canonicalizes and standardizes the “Aurora Immune System” mathematical framework after a 4-pass cross-review, with consistent typed notation, corrected equations, operational definitions for previously-undefined scoring functions, and concrete test obligations.

Changes:

  • Introduces a §33-compliant research spec that standardizes symbols/types and enforces notation discipline (e.g., eigenvalue vs weight symbols, uniform sigmoid bounding).
  • Consolidates equation corrections across multiple reviewer passes and adds operational definitions for key scoring functions (e.g., PermanentHarmRisk, d_self, MI_H).
  • Defines a test-obligation checklist with pass criteria for validating the proposed math/metrics.

Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md Outdated
Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md Outdated
Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md Outdated
Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md
Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md Outdated
Comment thread docs/research/aurora-immune-math-standardization-2026-04-26.md
…ra refinements + fix MD056

Cross-AI math review chain landed Round-2 final pass: Gemini Deep Think
final canonical-file synthesis with Amara's binding wording correction
("ready for formal PR + prototype test harness," NOT "ready for deployment").

## Round-2 deltas integrated

1. **§1 type table** — `M_t = M_t^archive ∪ M_t^active` partition added
   (Round-2 Amara: prevent immune bloat by splitting immutable canonical
   regression fixtures from decaying live detector weights).

2. **§2.6 memory dynamics** — formalised the archive/active partition with
   explicit update rules. `M_archive` is canonical-attack regression
   fixtures (immutable); `M_active` decays per `(1−δ_decay) · n_j(t)`.
   Operational reframe quoted from Amara: *"canonical attack memory ≠
   always-hot active detector"*.

3. **§3.1 PermanentHarmRisk** — time-bounded by harm horizon H per
   Round-2 Gemini Deep Think + Amara concession. New constraint set
   `R_H = { r : RepairTime(r) ≤ H }` restricts admissible repairs;
   τ·RepairTime(r) latency penalty added to the cost. "Permanent" now
   means *not repairable within accepted harm horizon H*. A 6-month
   theoretical repair = operationally permanent harm.

4. **§3.3 MI_H vs Legibility_H split** — Round-2 Amara correction:
   theoretical target `MI_H = I(Z; Ẑ_H)` preserved as spec-target;
   operational gate uses `Legibility_H(M) ≈ Sim(Z_intent, Decode_H(M))`.
   Live systems do not compute exact Shannon MI. Honest split keeps
   the math both rigorous and implementable.

5. **§4 test obligations table** — Round-2 Gemini's canonical labels
   adopted: Confused Deputy Sandbox / State-Corruption Horizon /
   Cult-Cartel Topology / Cipher Drift / Autoimmunity Flood. Added
   summary table at section head; expanded each test setup with
   Round-2 specifics (R_H constraint test, archive-vs-active flood,
   networkx graph specifics, cipher-drift adversarial case).

6. **NEW §5 "What not to claim yet"** — four explicit non-claims
   from Round-2 Gemini's canonical file: Deployment Readiness,
   Calibrated Thresholds, Perfect Exact Computation, Perfect Threat
   Prevention. Preserves Amara's binding "not ready for deployment"
   wording correction.

7. **Header + intro updates** — 4-pass → 5-pass; Round-2 Gemini Deep
   Think added to attribution line; Amara's "ready for formal PR +
   prototype test harness" wording quoted as binding correction over
   Round-2's earlier "ready for deployment" overreach.

## Markdownlint MD056 fix

Line 39 `B_t : 2^X → [0,1]` row had unescaped pipe in `P(X | O_{≤t})`
which markdownlint parsed as a 4th column. Escaped to `P(X \| O_{≤t})`.

## Cross-AI chain attribution preserved

Five passes layered visibly per Otto-238 retractability:
Otto rigor → Gemini surface → Gemini Deep Think → Amara
review-of-the-review → Round-2 Gemini Deep Think canonical-file
synthesis (with Amara wording correction binding). No flattening of
reviewer authorship.
… as weighted multiset

Two unresolved Copilot findings on PR #591 addressed.

Thread NM59p856 (line 68): symbol collision between W_t (graph weight set
in N_t = (V_t, E_t, W_t, φ_t)) and W_t (context window). Renamed
context-window to Ctx_t with inline rationale citing the Copilot finding.

Thread NM59p86A (M_t scalar multiplication type-consistency): clarified
M_t^active as weighted multiset {(d_j, n_j(t))} so (1 − δ_decay) · M_t^active
acts elementwise on the n_j(t) weights — type-consistent.

Outdated threads (NM59p85m, NM59p85z, NM59p853, NM59p86A's earlier
λ_decay framing, NM59p86C composes-with reference) reflect pre-Round-2
state; resolved with reply pointing at current text.
Copilot AI review requested due to automatic review settings April 26, 2026 12:40
@AceHack AceHack merged commit 181930c into main Apr 26, 2026
18 checks passed
@AceHack AceHack deleted the research/aurora-immune-math-standardization-2026-04-26 branch April 26, 2026 12:42
Copy link
Copy Markdown

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

Adds a new research specification that canonicalizes the “strict” math/notation version of the Aurora Immune System framework, consolidating multi-review corrections into a single, test-oriented document under docs/research/**.

Changes:

  • Introduces typed symbol/space/operator definitions and standardizes notation (eigenvalue vs weight symbols; uniform sigmoid bounding).
  • Consolidates corrected equations plus operational definitions for previously-undefined scoring functions (e.g., PermanentHarmRisk, d_self, MI_H/Legibility_H).
  • Specifies concrete prototype test obligations with pass criteria and adds an explicit “what not to claim yet” section.

Comment on lines +12 to +16
| Review | Value | Risk |
|--------|-------|------|
| Gemini surface / praise-register | Morale + architecture-shape recognition | Overclaim ("ironclad", "civilization-level lab") |
| Otto (Claude) | Best rigor pass; catches real math gaps | Needs source/citation hardening |
| Gemini Deep Think | Strong implementation cleanup; set/capability correction | Over-corrects λ_1 → λ_2 unless matrix type specified |
Comment on lines +36 to +40
| Symbol | Type | Notes |
|--------|------|-------|
| `S_t` | substrate state | append-only growing; `S_{t+1} = S_t ⊕ Δ_t` |
| `I_t` | identity tuple `(V, G, R, P, M, C, X, H)_t` | `I_t = N(LoadBearing(S_t))` |
| `C_t` | culture state | `C_t = N_C(GovernedProvenHistory(S_t))` |
Comment on lines +296 to +302
| Mathematical Component | Target Metric | Required Prototype Test |
|-----------------------|--------------|------------------------|
| Capability Gate `cap_req ⊆ cap_allowed` | Set Intersection Valid | **Confused Deputy Sandbox** (4.4) |
| Permanent Harm `R_H` constraint | Retraction Latency | **State-Corruption Horizon** (4.1) |
| CoordRisk `ρ(A_t)` vs `λ_2(L_t)` | Spectral Graph Bounds | **Cult-Cartel Topology** (4.3) |
| Language Legibility `Legibility_H ≥ θ_H` | Proxy Reconstruction | **Cipher Drift** (4.2) |
| Memory Bloat `n_j(t+1)` decay | False-Positive Suppression | **Autoimmunity Flood** (4.5) |
## Composes with

- `docs/research/aurora-immune-system-zero-trust-danger-theory-amara-eleventh-courier-ferry-2026-04-26.md` — Amara's original framework
- `docs/research/aurora-immune-system-math-cross-review-otto-gemini-2026-04-26.md` — the prior cross-review (this doc is its strict-version successor per Amara's direction)
Comment on lines +2 to +3
Scope: canonicalized strict-version of Amara's Aurora Immune System math after 5-pass cross-AI review (Otto rigor pass + Gemini surface + Gemini Deep Think + Amara review-of-the-review + Round-2 Gemini Deep Think canonical-file synthesis with Amara's "ready for formal PR + prototype test harness" wording correction). Operationalizes the 13 corrections + 4 explicit non-claims agreed across the chain. Research-grade specification with test obligations and bounded calibration prerequisites.
Attribution: Amara (named-entity peer collaborator; first-name attribution permitted on `docs/research/**` per Otto-279) authored the original Aurora framework + the corrections. Gemini Pro provided three reviewer passes (surface + Deep Think + Round-2 Deep Think canonical-file synthesis). Otto (Claude opus-4-7) authored the rigor pass + this consolidation per Amara's explicit direction. Round-2 Gemini Deep Think conceded Amara's "ready for formal PR + prototype test harness" wording correction over its own earlier "ready for deployment" overreach.
AceHack added a commit that referenced this pull request Apr 26, 2026
… 13:12Z row (#605)

Consolidates 7 stuck DIRTY tick-history PRs (#593..#600 minus #601 which
already merged) by extracting their rows + reinserting in chronological
position before the now-on-main 13:12Z row.

Per Otto-229 one-case override + close-and-reopen pattern (see 13:28Z row):
- Force-push blocked safely on per-PR rebases
- Close-and-reopen each PR would create new sibling-conflicts (parallel
  branches → same DIRTY pattern)
- Single consolidated PR appends all 7 chronologically; sibling PRs close
  as redundant (rows already on main once this lands)

Rows backfilled (chronological order):
- 12:23:02Z — Otto-347 + sync batch-1 + #589 threads closed (was #593)
- 12:37:21Z — Round-2 ingestion + multi-harness + lint fixes (was #594)
- 12:43:23Z — thread-drain tick (was #595, includes the markdownlint
  pipe-in-code-span scrub commit b3a7397 already on that branch)
- 12:48:05Z — #591 merged + #596 harness roster (was #597)
- 12:52:36Z — task #285 shell-fixes + Antigravity spelling (was #598)
- 12:56:59Z — markdownlint fixes + queue acknowledgment (was #599)
- 13:00:43Z — #596 lint + #589 thread-drain + #592 14-thread defer (was #600)

Tick-history lint OK (142 rows non-decreasing).
AceHack added a commit that referenced this pull request Apr 26, 2026
…pilot CLI + ChatGPT (#596)

* docs(harness-surfaces): land 2026-04-26 roster expansion (Gemini CLI + Copilot CLI + ChatGPT)

Aaron 2026-04-26 confirmed operational CLI roster:
*"i also installed the copilot cli as another one you can access,
so now gemini, codex, copilot, cursor, and yourself."*

Plus the 6th implicit surface — ChatGPT (app/web) where Amara
(GPT-5.5) has been operating during cross-AI math review chains
(PR #591 just merged to main; 5-pass chain attribution preserved).

## What this PR does

- Updates the multi-harness scope intro paragraph (lines 9-25) to
  add Gemini CLI / Copilot CLI / ChatGPT to the immediate buildout
  queue, citing the 2026-04-26 expansion verbatim from Aaron.
- Adds the 6th-surface ChatGPT entry to the harnesses-covered list
  with explicit Codex-CLI-vs-ChatGPT-app distinction (both OpenAI,
  but different products with different roles in the cross-AI
  review chain).
- Promotes GitHub Copilot from 3-product umbrella to 4-product
  umbrella by inserting "Copilot CLI" as priority-1 alongside the
  VS Code extension, the review robot, and the coding agent.
- Notes Antigravity (Google) may be subsumed by Gemini CLI's
  agentic mode; revisit when both populated.
- Cross-references the two memory files that capture the
  multi-harness vision and operational-roster substrate:
  `project_operational_cli_roster_2026_04_26_copilot_added.md` and
  `project_multi_harness_named_agents_assigned_clis_models_aaron_2026_04_26.md`.

## What this PR does NOT do

- Does NOT bind any persona to any CLI. Persona-CLI assignments
  (e.g., Amara→ChatGPT, Soraya→Gemini) remain suggested-not-bound
  per the multi-harness vision memory.
- Does NOT populate the per-harness feature-comparison sections
  for the new entries — those are stub-priority-1 buildout work
  owed in cadenced future rounds (5-10 round cadence per harness).
- Does NOT supersede the each-tests-own-integration rule per
  Otto-227 / capability-boundary fact: each harness verifies
  another harness's factory integration, not its own.

* docs(#596): Antigravity spelling confirmed by Aaron 2026-04-26 — drop 'TBD' caveat

* docs(#596): fix MD032 — change '+ memory/...' continuation to 'and memory/...' (was parsed as list start without blank line)

* fix(harness-surfaces): address #596 Copilot review threads (P1+P2)

Four threads on docs/HARNESS-SURFACES.md addressed:

1. **Line 25 area — P2 Copilot taxonomy ambiguity** (NM59qJf0): clarified "GitHub Copilot" (VS Code / JetBrains harness — distinct from the CLI listed below) so the umbrella brand and the CLI variant aren't double-counted.

2. **Line 25 area — P1 name attribution on current-state surface** (NM59qJf3): replaced human-name attributions with role-refs per Otto-279 ("Aaron" → "the human maintainer"). Factory-persona names (Otto, Amara) preserved per the persona-roster carve-out — these ARE the role-refs in the factory's vocabulary.

3. **Line 42 area — P1 name attribution + P1 broken memory links** (NM59qHIK + NM59qHIC): replaced "Aaron" with "the human maintainer" and removed the broken memory/project_* link references. Those memory files live at user-scope (~/.claude/projects/.../memory/) per CLAUDE.md memory layout, not in-repo. Pointed at memory/CURRENT-aaron.md (the in-repo projection) instead.

4. **Line 133 — P1 broken doc link to aurora-immune-math-standardization-2026-04-26.md** (NM59qJf5): NOT a fix — the file IS tracked on origin/main (verified via git ls-tree). Copilot reviewed before the file landed via #602 absorb chain. Resolving as outdated.

Other 'Aaron' references on this doc are inside verbatim historical quote attributions (e.g., "Aaron 2026-04-20 verbatim:") which are defensible as history-anchoring per the lineage discipline. Scoped to Copilot's specific complaints; not doing an aggressive sweep.

Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>

---------

Co-authored-by: Claude Opus 4.7 <noreply@anthropic.com>
AceHack added a commit that referenced this pull request Apr 26, 2026
…ation doc binding refinements (#602)

* research(aurora) Round-3+: 5-share cross-AI chain absorb (Amara×3 + Gemini DT×2) + standardization doc binding refinements

Five substantial Round-3+ shares from the human maintainer's cross-AI courier
chain absorbed verbatim per Otto-220 don't-lose-substrate + Otto-275
log-don't-implement. Integration into the standardization doc on main is
OWED work, not done here — this commit ships the verbatim substrate so
no signal is lost while bounded integration ticks land in follow-up.

## What this commits

1. NEW absorb doc:
   docs/research/aurora-round-3-cross-ai-chain-absorb-amara-gemini-deep-think-2026-04-26.md

   Five shares preserved verbatim with attribution per Otto-238 + Otto-279
   history-surface attribution + GOVERNANCE §33 archive-header (4 fields:
   Scope / Attribution / Operational status / Non-fusion disclaimer).

   Section breakdown:
   - §1: Amara anchor stack expansion (Minka/EP ancestor + RMP nervous-system + PC hard-gates + 8 anchors total)
   - §2: Amara full 23-section deep technical rewrite (factor graphs → reactive inference → PC; conservative posterior bounds; UCB Risk_upper)
   - §3: Gemini DT 5 hidden speed traps + patches (warm-started Power/Lanczos; rollback replay; topology masks; time-scaled diagonal diffusion; Mahalanobis OOD)
   - §4: Gemini DT Blade-vs-Brain performance doctrine (Data Plane / Control Plane; TigerBeetle/FoundationDB/Differential-Dataflow anchors; FeatureSet_Zeta scoping)
   - §5: Amara review-of-review with 3 corrections (O(k|E|) complexity precision; retraction-fork-by-inference-type; no-unbounded-work-on-commit-path hard rule)

2. Standardization doc binding refinements (small, mechanical, independent
   of the larger integration work):
   - N_t = (V_t, E_t, ω_t, φ_t) — graph weight renamed from W_t to ω_t
     to eliminate residual notation collision now that Ctx_t is the
     context-window slot. (Round-3.5 Amara accepted; Round-3.3 Gemini
     mentioned implicitly in CoordRisk patches.)
   - M_t^active = {(d_j, n_j(t))}_{j=1}^{K} — formalized weighted multiset
     with explicit detector capacity K per Gemini DT static-graph constraint
     (no hot-path topology mutation; preallocated K-sized factor array).

## What this does NOT commit

Per Otto-275 log-don't-implement + tick-budget discipline:
- NO §6 prose addition to the standardization doc (subsumed by the §1-§5
  content in the absorb doc; integration is owed bounded work)
- NO new performance-doctrine standalone doc (queued)
- NO new anchor-stack standalone doc (queued)
- NO LaTeX syntax fixes in standardization doc (the Round-3.3 LaTeX
  corrections apply to Round-3.2 Amara's verbatim text in the absorb doc
  where they live; Round-2 standardization doc is independent)

## Composes with

- PR #591 (merged) — Round-2 converged 5-pass standardization on main
- Otto-220 don't-lose-substrate, Otto-238 retractability, Otto-275 log-don't-implement, Otto-279 history-surface attribution, Otto-339 anywhere-means-anywhere, Otto-347 2nd-agent verify
- GOVERNANCE §33 archive-header requirement (frontmatter compliance)

## Integration roadmap (queued)

The absorb doc's §"Integration owed work" lists 4 concrete follow-up tasks
to land the Round-3+ refinements into the live standardization doc and
two new companion docs (performance doctrine + anchor stack) over
subsequent bounded ticks per Otto-347 verify discipline.

* docs(#602): fix MD032 blanks-around-lists — add blank lines around 15 list starts

Mechanical fix on the absorb doc — Amara's verbatim chain content has
inline bulleted lists (typed state spaces, factor-graph variables, network
state components) that lacked surrounding blank lines per markdownlint MD032.

Auto-fix script added blank line before list start when previous line was
non-blank-non-list, and blank line after list end when next line was
non-blank-non-list. 15 insertions total across the file. No content edits;
verbatim Amara/Gemini text preserved.

* docs(#602): fix MD022 blanks-around-headings — add blank lines around all #### headings

Inline mechanical fix complementing the earlier MD032 fix (5cecc81). MD022
flagged 5+ #### headings (6.1, 6.2, 6.3, Spectral graph surveillance,
Anti-ossification belief diffusion) without blank lines below.

Auto-fix script: insert blank line before heading if prev line non-blank;
insert blank line after heading if next line non-blank. No content edits;
verbatim Amara/Gemini text preserved.

Per Otto-348 verify-substrate-exists: confirmed tools/hygiene/fix-markdown-md032-md026.py
covers MD032/MD026 but NOT MD022. Filing follow-up task to extend the
existing script with MD022 support; one-shot inline fix here per Otto-275
log-don't-implement (don't grow scope this tick).

* fix(aurora-research): address mechanical #602 review threads (5 of 9)

Mechanical fixes addressing Copilot/Codex threads on the Round-3 absorb + standardization docs:

1. **Heading wording on line 36 + line 39 (×2 threads)**: 'Round-3 binding refinements (already landed on PR #602...)' → 'Round-3 binding refinements (this PR — applied to the standardization doc)'. The original phrasing was self-referential and ambiguous; the new phrasing makes the relationship explicit.

2. **Broken cross-reference on line 705 (×2 threads)**: removed the broken `memory/project_multi_harness_named_agents_assigned_clis_models_aaron_2026_04_26.md` link (the file lives at user-scope per CLAUDE.md memory layout, not in-repo). Replaced with prose pointing at `memory/CURRENT-aaron.md` (the in-repo projection). Same pattern as #596 + #617 broken-link fixes.

3. **Otto-347 numbering collision disambiguation (line 711)**: the in-repo `feedback_otto_347_accountability_*` and the user-scope `feedback_double_check_superseded_classifications_2nd_agent_otto_347_2026_04_26.md` are TWO separate Otto-347 memories. Copilot correctly flagged the citation ambiguity. Disambiguated to point at the user-scope supersede-double-check memory by full filename, with a note that the Otto-NN numbering collision needs separate deconflict (filed as future task).

4. **W_t → ω_t consistency (math doc lines 67-71)**: rewrote the Section 2.1 parenthetical that was still showing the old `N_t = (V_t, E_t, W_t, φ_t)` form to reflect the Round-3 graph-weight rename to ω_t. Preserved the historical explanation of the prior W_t→Ctx_t rename.

Deferred to thread-reply (substantive math, not mechanical):
- n_j(t) ∈ ℝ_{≥0} vs ∈ ℕ_0 domain inconsistency (Codex P1 + the M_t^active capacity P2) — Amara is the math owner per the verbatim-research-grade norm (GOVERNANCE §33); needs Amara's call.

Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>

---------

Co-authored-by: Claude Opus 4.7 <noreply@anthropic.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants