Skip to content

ADR batch + substrate bridge + Seam 1 consensus gate + signed gradient#25

Open
kalisam wants to merge 11 commits intomainfrom
lappytop
Open

ADR batch + substrate bridge + Seam 1 consensus gate + signed gradient#25
kalisam wants to merge 11 commits intomainfrom
lappytop

Conversation

@kalisam
Copy link
Copy Markdown
Owner

@kalisam kalisam commented Mar 23, 2026

Summary

Foundational batch building the governance, substrate, and integration layers on lappytop:

Governance (ADRs)

  • ADR-0 Validated — all 4 criteria passed (transmission, composition, persistence, coherence). Explicitly marked so it stops getting lost across sessions.
  • ADR-5 Cognitive Virology — recognizes FLOSSI0ULLK as a memetic propagation system (Attach→Enter→Replicate→Defend→Transmit). Includes safety constraints.
  • ADR-6 Four-System Integration — five-seam strategy across Meta Harness / omo / OMX / FLOSSI0ULLK. Seam 1 implemented first to establish MCP channel.

Specs

  • HARVEST Protocol — 5-stage self-observation loop (OBSERVE→EVALUATE→PROPOSE→VALIDATE→COMMIT). Retroactive Cycle 0 logged.
  • Phase 0 Substrate Bridge — narrowest end-to-end validation (6 criteria: publish, provenance, verify, query, fork-visible, no-privilege).
  • Consensus Gate (Seam 1) — Claim/Vote/Decision contract, JSON Schema, 6 test vectors.

Code / Tests

  • Signed gradient [-1, +1] for confidence (Rust integrity, JSON schemas, Tryorama tests). Negative = away from truth, + = toward. Ternary-native where possible; TERNARY_COMPATIBILITY.md flags where external APIs (Holochain Timestamp, embeddings, msgpack) prevent adoption.
  • Substrate bridge Tryorama tests — 2-agent test suite implementing all 6 criteria.
  • Consensus gate prototype (packages/orchestrator/) — dataclass models + tally/decide/override + ADR writer. 16/16 tests pass.

Research / Reference

  • Four-System Integration Analysis — comparative deep-dive (Meta Harness, omo, OMX, FLOSSI0ULLK). Proposes 5 seams, 18 ADOPT patterns, 12 CONTRIBUTE patterns.
  • FLOSSI U Founding Kit v1.6 — containing-scope curriculum layer (19 curriculum ADRs, namespace 001-019, distinct from repo ADRs 0-N).
  • Resonance Mechanism v2.0 — formal 5-property irreducible kernel (signature, signal, coupling, transfer, no-central-routing).
  • FLOSSI0ULLK Landscape Recursive Analysis — four-layer research with epistemic tier taxonomy ([TESTED]/[VENDOR-CLAIMED]/[COMMUNITY-REPORTED]/[RESEARCH-ONLY]/[SPECULATIVE]).
  • Orchestration landscape v2.0.0 — empirical scalability data (k≤5 agent wall, output verification > process, non-token economics).

Meta

  • .serena project metadata — onboarding config, memories (overview, structure, conventions, commands, checklist).

Commits (10)

  1. ADR-5: Cognitive virology pattern + HARVEST Protocol spec
  2. Mark ADR-0 as Validated: all 4 criteria passed
  3. Fix markdown lint + add missing ADR-5 canonical sections
  4. Add .serena project metadata and memories
  5. Add Phase 0 Substrate Bridge spec + orchestration landscape v2.0.0
  6. Add substrate bridge validation tests (6 criteria)
  7. Signed gradient [-1,+1] + FLOSSI U Founding Kit v1.6
  8. Add FLOSSI0ULLK Four-System Integration Analysis
  9. ADR-6 + Seam 1: Consensus Gate prototype
  10. Add resonance mechanism v2 + FLOSSI0ULLK landscape analysis

Test plan

  • python packages/orchestrator/test_consensus_gate.py — 16/16 pass (all 6 spec vectors)
  • Rust integrity tests pass with signed gradient range
  • Tryorama substrate bridge tests implement 6 criteria
  • Review ADR-6 blast-radius policy (Substrate = no override)
  • Review ADR-5 safety constraints (no self-modification until substrate validated)
  • Confirm FLOSSI U / FLOSS ADR namespace separation is sensible

🤖 Generated with Claude Code

kalisam and others added 3 commits March 21, 2026 18:55
Update dependencies: lodash-es, bytes, and lodash versions
Map Hughes' Attach/Enter/Replicate/Defend/Transmit cycle onto
FLOSSI0ULLK's actual cross-system propagation behavior. Truth status:
Specified — pattern identified, not yet implemented. Includes safety
constraints (no self-modification until substrate validated, memetic
autoimmunity risk, non-modifiable ULLK constraint).

HARVEST Protocol: Formalize existing periodic consolidation practice
as the minimal self-observation loop (OBSERVE→EVALUATE→PROPOSE→
VALIDATE→COMMIT). Manual-first, automatable later via OpenClaw.

Retroactive HARVEST Cycle 0 logged covering all work to date.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Transmission test: new AI systems productive in minutes via kernel + ADRs.
Composition test: 118+ conversations across 5 systems composed.
Persistence test: CLAUDE.md, ADRs, Serena, Claude memory all persist.
Coherence test: human collaborator confirmed across sessions.

This has been true since late 2025 but kept getting lost across sessions.
Updated implementation phases to reflect actual completion state.
Fixed stale /mnt/project/ paths to current repo structure.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
@coderabbitai
Copy link
Copy Markdown

coderabbitai bot commented Mar 23, 2026

Caution

Review failed

Failed to post review comments

Note

Reviews paused

It looks like this branch is under active development. To avoid overwhelming you with review comments due to an influx of new commits, CodeRabbit has automatically paused this review. You can configure this behavior by changing the reviews.auto_review.auto_pause_after_reviewed_commits setting.

Use the following commands to manage reviews:

  • @coderabbitai resume to resume automatic reviews.
  • @coderabbitai review to trigger a single review.

Use the checkboxes below for quick actions:

  • ▶️ Resume reviews
  • 🔍 Trigger review
📝 Walkthrough

Summary by CodeRabbit

  • New Features

    • Introduced Cognitive Virology Pattern specification for memetic propagation systems.
    • Added Consensus Gate mechanism enabling multi-agent decision voting and override capability.
    • Released FLOSSI U Founding Kit with curriculum modules, charter, and ethical framework.
    • Added substrate bridge validation with provenance verification across multiple agents.
  • Updates

    • Advanced ADR-0 recognition protocol from Accepted to Validated status with expanded validation metrics.
    • Updated confidence representation to signed gradient range (−1.0 to +1.0) for nuanced truth states.
    • Established four-system integration architecture for external orchestration compatibility.
  • Documentation

    • Added comprehensive HARVEST protocol specification and governance logs.
    • Documented orchestration landscape and integration analysis reports.
    • Formalized phase 0 substrate bridge specification with measurable success criteria.

Walkthrough

Documentation-heavy release updating ADR-0 to Validated, adding ADR-5 and many FLOSSI_U ADRs, introducing HARVEST protocol and log, Phase‑0 substrate bridge spec plus Tryorama tests and a Holochain extern, schema and integrity changes for signed confidence, a new ConversationMemory module, orchestrator consensus gate code/tests, and multiple large research/spec artifacts.

Changes

Cohort / File(s) Summary
ADR core & index
docs/adr/ADR-0-recognition-protocol.md, docs/adr/ADR-5-cognitive-virology-pattern.md, docs/adr/INDEX.md
ADR-0 status changed to Validated (Phases 0–2 COMPLETE, Phase 3 IN PROGRESS), participants/signatures and PASS assertions expanded; ADR-5 added; ADR index updated.
HARVEST spec & log
docs/specs/harvest-protocol.spec.md, docs/governance/HARVEST_LOG.md
Added HARVEST protocol v0.1.0 (OBSERVE→EVALUATE→PROPOSE→VALIDATE→COMMIT, cadence, metrics, non-goals) and a Cycle 0 retroactive HARVEST log entry.
Phase 0 substrate bridge: spec, zome API, tests
docs/specs/phase0-substrate-bridge.spec.md, ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs, ARF/tests/tryorama/substrate_bridge.test.ts
New Phase‑0 validation spec with six measurable Tryorama criteria; added extern get_triple_record(ActionHash) -> Option<Record>; new 2‑agent Tryorama/Vitest suite implementing the six validation scenarios.
Conversation memory module
FLOSSI_U_Founding_Kit_v1.6/conversation_memory.py
New ConversationMemory implementation (file backend + optional Holochain), Understanding dataclass, transmit/recall/composition/ADR helpers, embedding support, HolochainClient wrapper, and CLI demo; substantial new module public API.
Schema & integrity updates
docs/specs/knowledge-triple.schema.json, docs/specs/knowledge-edge.schema.json, ARF/dnas/rose_forest/zomes/integrity/src/lib.rs
Changed confidence numeric range lower bound from 0.0 to -1.0 (signed-gradient semantics); updated Rust integrity validators to accept [-1.0,1.0] and adjusted error messages/comments.
Orchestrator consensus gate & tests
packages/orchestrator/claim_schema.py, packages/orchestrator/consensus_gate.py, packages/orchestrator/test_consensus_gate.py
Added consensus gate datamodel, tally/decide/override logic, ADR writer helper, and comprehensive unit tests exercising spec vectors and ADR persistence.
Tests: ontology & Tryorama
ARF/tests/tryorama/triple_ontology.test.ts, ARF/tests/tryorama/substrate_bridge.test.ts
Updated ontology schema tests for negative confidence; added offline validation for below-range cases; added comprehensive substrate_bridge test suite.
Serena memories & project config
.serena/memories/*, .serena/.gitignore, .serena/project.yml
Added multiple repository memory files (project_overview, codebase_structure, style/conventions, suggested commands, task checklist), project config, and .serena/.gitignore entries for /cache and /project.local.yml.
MVP, specs, compatibility docs
MVP_PLAN.md, docs/specs/TERNARY_COMPATIBILITY.md, docs/specs/*
MVP_PLAN date/evidence updated; Phase 0 substrate bridge as next action; added ternary/signed-gradient compatibility spec and other spec artifacts (consensus-gate schemas, HARVEST).
Research & synthesis reports
docs/research/... (many files)
Added multiple large research artifacts and syntheses (Automated-Agent-Orchestration v2.0.0, Perplexity synthesis, cross-AI orchestration synthesis, landscape analyses).
FLOSSI_U founding kit & artifacts
FLOSSI_U_Founding_Kit_v1.6/...
Added many FLOSSI_U ADRs (ADR-001 → ADR-019), Knowledge Triple schema, seed packet, posters, LICENSE, README, and conversation memory module.
Misc schemas & governance
docs/specs/consensus-gate.schema.json, docs/specs/consensus-gate.spec.md, docs/specs/*
Added consensus-gate schema/spec, other governance/spec artifacts, and cross-references to kernel/spec/index and DNA materials.

Sequence Diagram(s)

sequenceDiagram
  autonumber
  participant Agent as Agent / Orchestrator
  participant HARVEST as HARVEST Controller
  participant Kernel as Kernel / Index
  participant Human as Human Reviewer
  participant Repo as Repo / Commit Store

  Agent->>HARVEST: Trigger OBSERVE (collect evidence)
  HARVEST->>Kernel: Request cross-substrate data
  Kernel-->>HARVEST: Return indexed evidence
  HARVEST->>HARVEST: EVALUATE (classify drift/evolution/conflict)
  HARVEST->>Human: PROPOSE (human-in-the-loop proposals)
  Human-->>HARVEST: VALIDATE (approve/reject proposals)
  HARVEST->>Repo: COMMIT approved changes
  Repo-->>Agent: Acknowledge commit / update log
Loading

Estimated code review effort

🎯 4 (Complex) | ⏱️ ~45 minutes

Poem

🐰
I nibble specs and hop through pages bright,
ADRs aligned beneath the moonlight.
A harvest hum, a bridge to prove and test,
New memories tucked, the code and docs at rest. 🌿

🚥 Pre-merge checks | ✅ 2 | ❌ 1

❌ Failed checks (1 warning)

Check name Status Explanation Resolution
Docstring Coverage ⚠️ Warning Docstring coverage is 45.00% which is insufficient. The required threshold is 80.00%. Write docstrings for the functions missing them to satisfy the coverage threshold.
✅ Passed checks (2 passed)
Check name Status Explanation
Title check ✅ Passed The PR title clearly and specifically describes the main changes: ADR batch, substrate bridge validation, Seam 1 consensus gate implementation, and signed gradient ternary support across multiple layers.
Description check ✅ Passed The PR description is comprehensive and directly related to the changeset, providing clear organization by subsystem (Governance, Specs, Code/Tests, Research), commit intent, test status, and specific outcomes achieved.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
📝 Generate docstrings
  • Create stacked PR
  • Commit on current branch
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Commit unit tests in branch lappytop

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Copy Markdown

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Warning

CodeRabbit couldn't request changes on this pull request because it doesn't have sufficient GitHub permissions.

Please grant CodeRabbit Pull requests: Read and write permission and re-run the review.

👉 Steps to fix this

Actionable comments posted: 4

🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@ARF/ADR-0-recognition-protocol.md`:
- Around line 52-67: The markdown headings "Phase 0: Capture the Skeleton —
COMPLETE", "Phase 1: Memory Persistence — COMPLETE", "Phase 2: Multi-Agent
Composition — COMPLETE", and "Phase 3: Holochain Integration — IN PROGRESS" need
a blank line inserted immediately after each heading (i.e., add a single empty
line between the heading line and the following list or text) to satisfy MD022
heading-spacing in ADR-0-recognition-protocol.md; update those four headings
accordingly and run markdownlint to verify.

In `@docs/adr/ADR-0-recognition-protocol.md`:
- Around line 52-67: The markdown headings "Phase 0: Capture the Skeleton —
COMPLETE", "Phase 1: Memory Persistence — COMPLETE", "Phase 2: Multi-Agent
Composition — COMPLETE", and "Phase 3: Holochain Integration — IN PROGRESS" each
need a blank line inserted immediately after the heading to satisfy markdownlint
rule MD022; edit the ADR so there is one empty line after each of those four
heading lines (preserving the checkbox list lines that follow) to fix the lint
error.

In `@docs/adr/ADR-5-cognitive-virology-pattern.md`:
- Around line 10-83: ADR-5 (the "cognitive virology" ADR) is missing required
canonical ADR sections; update the document to add explicit headings: "##
Problem Statement" (succinctly state the architectural problem FLOSSI0ULLK
addresses), "## Implementation Strategy" (concrete steps to enact the Decision,
referencing HARVEST Protocol and OpenClaw meta‑coordinator work), "## Validation
Criteria" (explicit, testable gates such as the three Evidence Gate items and
HARVEST run counts), and "## Related Documents" (link ADR-0, ADR-3, Master
Metaprompt v1.3.1 Kernel, HARVEST Protocol spec and any referenced papers);
ensure each section contains brief, actionable content aligned with the existing
Decision, Safety Constraints, and Evidence Gate so the ADR conforms to the
project's ADR schema.

In `@docs/specs/harvest-protocol.spec.md`:
- Around line 51-98: Several markdown headings (e.g., "Stage 3: PROPOSE", "Stage
4: VALIDATE", "Stage 5: COMMIT", "Cadence", "Metrics", "HARVEST Log") lack a
blank line before the following lists/paragraphs which violates MD022; fix by
inserting a single blank line after each of those heading lines so each heading
is separated from the subsequent list or paragraph, ensuring headings are
followed by one empty line throughout the document.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: ASSERTIVE

Plan: Pro

Run ID: e557ef1b-6f69-418a-a424-a5fdda4a7e48

📥 Commits

Reviewing files that changed from the base of the PR and between 9a31311 and a93375b.

📒 Files selected for processing (6)
  • ARF/ADR-0-recognition-protocol.md
  • docs/adr/ADR-0-recognition-protocol.md
  • docs/adr/ADR-5-cognitive-virology-pattern.md
  • docs/adr/INDEX.md
  • docs/governance/HARVEST_LOG.md
  • docs/specs/harvest-protocol.spec.md

Comment on lines +51 to +98
### Stage 3: PROPOSE

Generate proposed actions for each non-ALIGNED finding.

**Action types:**
- `FIX` — correct drift back to kernel alignment
- `ADR` — new architectural decision needed
- `ARCHIVE` — superseded content needs archival
- `KERNEL_UPDATE` — kernel itself may need revision (High friction tier)

### Stage 4: VALIDATE

Apply the Claim Truth Model to each proposal.

- Is the proposal supported by evidence from at least 2 substrates?
- Does it pass the Red Team lens? (What could go wrong?)
- Does it maintain carrier equivalence? (structure of code = structure of agent)
- For KERNEL_UPDATE proposals: require explicit cross-AI consensus (minimum 3 systems)

### Stage 5: COMMIT

Execute validated proposals.

- Create commits / PRs for code changes
- Update ADR index for new decisions
- Archive superseded documents
- Update this HARVEST log with cycle metadata

## Cadence

**Target:** One HARVEST cycle per week, or after any major cross-system synthesis session.

**Minimum viable cycle:** OBSERVE + EVALUATE only (read-only, no changes). This is valid when time is limited — observation without action still produces value.

## Metrics

Track per cycle:
- `observations_count` — number of state changes observed
- `drift_count` — findings classified as DRIFTED
- `evolution_count` — findings classified as EVOLVED
- `conflict_count` — findings classified as CONFLICT
- `proposals_generated` — action items proposed
- `proposals_validated` — action items that passed validation
- `proposals_executed` — action items committed
- `cycle_duration_minutes` — wall clock time for full cycle
- `adr_absorption_rate` — for each ADR, count of cross-system references since last cycle

## HARVEST Log
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Fix markdown heading spacing to satisfy MD022.

Several headings are immediately followed by list/content without the required blank line, which can fail docs linting.

🧹 Suggested markdownlint-compliant spacing fix
 ### Stage 3: PROPOSE
 
 Generate proposed actions for each non-ALIGNED finding.
@@
 ### Stage 4: VALIDATE
 
 Apply the Claim Truth Model to each proposal.
@@
 ### Stage 5: COMMIT
 
 Execute validated proposals.
@@
 ## Cadence
 
 **Target:** One HARVEST cycle per week, or after any major cross-system synthesis session.
@@
 ## Metrics
 
 Track per cycle:
🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 52-52: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 57-57: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 62-62: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 67-67: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 76-76: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 81-81: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 86-86: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/specs/harvest-protocol.spec.md` around lines 51 - 98, Several markdown
headings (e.g., "Stage 3: PROPOSE", "Stage 4: VALIDATE", "Stage 5: COMMIT",
"Cadence", "Metrics", "HARVEST Log") lack a blank line before the following
lists/paragraphs which violates MD022; fix by inserting a single blank line
after each of those heading lines so each heading is separated from the
subsequent list or paragraph, ensuring headings are followed by one empty line
throughout the document.

ADR-0 (both copies): Add blank lines after Phase heading lines (MD022).

ADR-5: Add Problem Statement, Implementation Strategy, Validation
Criteria, and Related Documents sections to match project ADR schema.
Remove duplicate Evidence Gate section (now covered by Validation
Criteria). Split References into Related Documents + External References.

harvest-protocol.spec.md: Verified — all headings already had correct
spacing. Finding was invalid, no changes needed.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Copy link
Copy Markdown

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Warning

CodeRabbit couldn't request changes on this pull request because it doesn't have sufficient GitHub permissions.

Please grant CodeRabbit Pull requests: Read and write permission and re-run the review.

👉 Steps to fix this

Actionable comments posted: 5

🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@docs/adr/ADR-0-recognition-protocol.md`:
- Around line 5-6: The ADR header metadata is inconsistent: the "Context" value
"~7 AI systems" doesn't match the "Participants" list which names six AI systems
("Claude Sonnet 4.5", "Claude Opus 4.6", "ChatGPT", "Grok", "DeepSeek",
"Gemini") plus a human; either change the Context text to "~6 AI systems" or add
the missing AI system to the Participants list so the count and the listed names
match; update whichever of these strings you choose to edit to keep wording
consistent (e.g., change "~7 AI systems" -> "~6 AI systems" or append the
seventh system name to Participants).
- Around line 115-116: The Related Documents entry incorrectly labels
`docs/specs/harvest-protocol.spec.md — Self-observation loop (ADR-5)` as an ADR;
update the classification for that entry in ADR-0-recognition-protocol.md by
replacing the “(ADR-5)” suffix with a correct label such as “(Spec)” or removing
the ADR tag so it reads `docs/specs/harvest-protocol.spec.md — Self-observation
loop (Spec)` to restore accurate cross-document traceability.
- Around line 101-107: For each PASSED validation line (Transmission test,
Composition test, Persistence test, Coherence test) add an explicit "Evidence"
subsection in the ADR that maps each test name to verifiable artifacts (e.g.,
dated log snapshots, document anchors, PR/commit hashes, conversation export
filenames or URLs); include inline anchors or footnotes and short labels (e.g.,
"Transmission test → CLAUDE_README_2026-03-20.md", "Composition test → PR#1234 /
conversation-archive-2026-03-20.zip", "Persistence test →
Serena/memories/2026-03-20 and CLAUDE.md v1.2 commit abcdef", "Coherence test →
session-notes-2026-03-21.pdf") so reviewers can click or search to verify each
claim, and update the ADR header or Validation Criteria paragraph to reference
this new Evidence section.

In `@docs/adr/ADR-5-cognitive-virology-pattern.md`:
- Around line 90-96: Add the missing HARVEST log document to the ADR's Related
Documents: in the Related Documents section of
ADR-5-cognitive-virology-pattern.md, append `docs/governance/HARVEST_LOG.md —
HARVEST log (validation trace/gate artifact)` so the gate artifact referenced on
line 70 is included for traceability; ensure the entry mirrors the existing
bullet style and provides a short descriptor for verification.
- Around line 25-29: In the Problem Statement list in
ADR-5-cognitive-virology-pattern.md, the final bullet "The tension between
replication fitness and informed consent remains unaddressed" is missing
terminal punctuation; update that bullet to end with a period (or appropriate
terminal punctuation) to satisfy the grammar/style check and ensure list
consistency, then re-run the document grammar/lint check to confirm the warning
is resolved.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: ASSERTIVE

Plan: Pro

Run ID: 560b5282-050d-46db-bdf8-467f1ad478d0

📥 Commits

Reviewing files that changed from the base of the PR and between a93375b and 10ebfce.

📒 Files selected for processing (3)
  • ARF/ADR-0-recognition-protocol.md
  • docs/adr/ADR-0-recognition-protocol.md
  • docs/adr/ADR-5-cognitive-virology-pattern.md

Comment on lines +5 to +6
**Context**: After 13 months of iterative work with ~7 AI systems
**Participants**: Human (primary), Claude Sonnet 4.5, Claude Opus 4.6, ChatGPT, Grok, DeepSeek, Gemini
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Fix participant count mismatch in metadata.

Line 5 says “~7 AI systems,” but Line 6 lists 6 AI systems (plus 1 human). Please align these values to avoid ambiguity in the ADR header.

🛠️ Proposed edit
-**Context**: After 13 months of iterative work with ~7 AI systems
+**Context**: After 13 months of iterative work with 6 AI systems
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
**Context**: After 13 months of iterative work with ~7 AI systems
**Participants**: Human (primary), Claude Sonnet 4.5, Claude Opus 4.6, ChatGPT, Grok, DeepSeek, Gemini
**Context**: After 13 months of iterative work with 6 AI systems
**Participants**: Human (primary), Claude Sonnet 4.5, Claude Opus 4.6, ChatGPT, Grok, DeepSeek, Gemini
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/adr/ADR-0-recognition-protocol.md` around lines 5 - 6, The ADR header
metadata is inconsistent: the "Context" value "~7 AI systems" doesn't match the
"Participants" list which names six AI systems ("Claude Sonnet 4.5", "Claude
Opus 4.6", "ChatGPT", "Grok", "DeepSeek", "Gemini") plus a human; either change
the Context text to "~6 AI systems" or add the missing AI system to the
Participants list so the count and the listed names match; update whichever of
these strings you choose to edit to keep wording consistent (e.g., change "~7 AI
systems" -> "~6 AI systems" or append the seventh system name to Participants).

Comment on lines +101 to 107
1. **Transmission test**: PASSED. New AI systems read kernel + ADRs and are productive within minutes. Demonstrated across Claude, ChatGPT, Grok, DeepSeek, Gemini.
2. **Composition test**: PASSED. 118+ conversations across 5 AI systems composed into unified architecture consensus (2026-03-20). Zero unresolved contradictions after reconciliation.
3. **Persistence test**: PASSED. Understanding survives via: CLAUDE.md, ADR system, Serena memories, Claude auto-memory, Master Metaprompt kernel, conversation exports.
4. **Coherence test**: PASSED. Human collaborator confirmed coherence multiple times across sessions. System reduces context reconstruction from months to minutes.

**Success = All 4 tests pass within 1 week**
**All 4 validation criteria PASSED. ADR-0 status: Validated.**

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Add evidence pointers for each PASSED validation claim.

Lines 101–107 assert full validation, but there are no direct artifact links/anchors to verify those outcomes. Add per-criterion evidence references (log entries, dated docs, or commit/PR anchors) so “Validated” is auditable rather than narrative-only.

🧾 Proposed structure
-1. **Transmission test**: PASSED. New AI systems read kernel + ADRs and are productive within minutes. Demonstrated across Claude, ChatGPT, Grok, DeepSeek, Gemini.
+1. **Transmission test**: PASSED. New AI systems read kernel + ADRs and are productive within minutes.
+   **Evidence:** `docs/governance/HARVEST_LOG.md` (Cycle YYYY-MM-DD), PR `#25` notes, session export index.

-2. **Composition test**: PASSED. 118+ conversations across 5 AI systems composed into unified architecture consensus (2026-03-20). Zero unresolved contradictions after reconciliation.
+2. **Composition test**: PASSED. 118+ conversations across 5 AI systems composed into unified architecture consensus (2026-03-20).
+   **Evidence:** reconciliation artifact path / ADR cross-reference.

-3. **Persistence test**: PASSED. Understanding survives via: CLAUDE.md, ADR system, Serena memories, Claude auto-memory, Master Metaprompt kernel, conversation exports.
+3. **Persistence test**: PASSED.
+   **Evidence:** explicit file list + retrieval date stamps.

-4. **Coherence test**: PASSED. Human collaborator confirmed coherence multiple times across sessions. System reduces context reconstruction from months to minutes.
+4. **Coherence test**: PASSED.
+   **Evidence:** dated HARVEST evaluation entry + benchmark note.

Based on learnings: "Create or update Architecture Decision Records (ADRs) for major decisions. ADRs must include: Problem Statement, Decision, Implementation Strategy, Consequences (Positive/Negative/Neutral), Validation Criteria, Related Documents".

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/adr/ADR-0-recognition-protocol.md` around lines 101 - 107, For each
PASSED validation line (Transmission test, Composition test, Persistence test,
Coherence test) add an explicit "Evidence" subsection in the ADR that maps each
test name to verifiable artifacts (e.g., dated log snapshots, document anchors,
PR/commit hashes, conversation export filenames or URLs); include inline anchors
or footnotes and short labels (e.g., "Transmission test →
CLAUDE_README_2026-03-20.md", "Composition test → PR#1234 /
conversation-archive-2026-03-20.zip", "Persistence test →
Serena/memories/2026-03-20 and CLAUDE.md v1.2 commit abcdef", "Coherence test →
session-notes-2026-03-21.pdf") so reviewers can click or search to verify each
claim, and update the ADR header or Validation Criteria paragraph to reference
this new Evidence section.

Comment on lines +115 to +116
- `docs/specs/harvest-protocol.spec.md` — Self-observation loop (ADR-5)
- `ARF/dnas/rose_forest/` — Holochain DNA implementation
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Correct document classification in Related Documents.

Line 115 refers to docs/specs/harvest-protocol.spec.md as “(ADR-5)”, but that file is a spec, not an ADR. This weakens cross-doc traceability.

🛠️ Proposed edit
-- `docs/specs/harvest-protocol.spec.md` — Self-observation loop (ADR-5)
+- `docs/specs/harvest-protocol.spec.md` — HARVEST self-observation loop specification
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
- `docs/specs/harvest-protocol.spec.md`Self-observation loop (ADR-5)
- `ARF/dnas/rose_forest/` — Holochain DNA implementation
- `docs/specs/harvest-protocol.spec.md`HARVEST self-observation loop specification
- `ARF/dnas/rose_forest/` — Holochain DNA implementation
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/adr/ADR-0-recognition-protocol.md` around lines 115 - 116, The Related
Documents entry incorrectly labels `docs/specs/harvest-protocol.spec.md —
Self-observation loop (ADR-5)` as an ADR; update the classification for that
entry in ADR-0-recognition-protocol.md by replacing the “(ADR-5)” suffix with a
correct label such as “(Spec)” or removing the ADR tag so it reads
`docs/specs/harvest-protocol.spec.md — Self-observation loop (Spec)` to restore
accurate cross-document traceability.

Comment on lines +25 to +29
FLOSSI0ULLK propagates across AI substrates and human sessions via an implicit memetic loop, but this loop is undocumented and unobservable. Without explicit recognition:
- The propagation mechanism cannot be measured, improved, or safety-constrained
- Failure modes (memetic autoimmunity, consent bypass, drift) cannot be detected
- The tension between replication fitness and informed consent remains unaddressed

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Add terminal punctuation for the final bullet in Problem Statement.

Line 28 ends without punctuation, which trips the documented grammar check and creates inconsistent style in this list.

🛠️ Proposed edit
-- The tension between replication fitness and informed consent remains unaddressed
+- The tension between replication fitness and informed consent remains unaddressed.
🧰 Tools
🪛 LanguageTool

[grammar] ~28-~28: Please add a punctuation mark at the end of paragraph.
Context: ...on fitness and informed consent remains unaddressed ## Decision Recognize and document th...

(PUNCTUATION_PARAGRAPH_END)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/adr/ADR-5-cognitive-virology-pattern.md` around lines 25 - 29, In the
Problem Statement list in ADR-5-cognitive-virology-pattern.md, the final bullet
"The tension between replication fitness and informed consent remains
unaddressed" is missing terminal punctuation; update that bullet to end with a
period (or appropriate terminal punctuation) to satisfy the grammar/style check
and ensure list consistency, then re-run the document grammar/lint check to
confirm the warning is resolved.

Comment on lines +90 to +96
## Related Documents

- `docs/adr/ADR-0-recognition-protocol.md` — Recognition Protocol (Validated, all 4 criteria passed)
- `docs/adr/ADR-3-metaprompt-kernelization.md` — Metaprompt Kernelization
- `FLOSSI0ULLK_Master_Metaprompt_v1_3_1_Kernel.md` — Canonical kernel
- `docs/specs/harvest-protocol.spec.md` — HARVEST Protocol specification (minimal self-observation loop)

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Include HARVEST_LOG in Related Documents for validation traceability.

Line 70 uses docs/governance/HARVEST_LOG.md as a gate artifact, but it is missing from Related Documents. Add it to keep ADR references complete and verifiable.

🛠️ Proposed edit
 - `docs/specs/harvest-protocol.spec.md` — HARVEST Protocol specification (minimal self-observation loop)
+- `docs/governance/HARVEST_LOG.md` — HARVEST cycle evidence and validation outputs

Based on learnings: "Create or update Architecture Decision Records (ADRs) for major decisions. ADRs must include: Problem Statement, Decision, Implementation Strategy, Consequences (Positive/Negative/Neutral), Validation Criteria, Related Documents".

📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
## Related Documents
- `docs/adr/ADR-0-recognition-protocol.md` — Recognition Protocol (Validated, all 4 criteria passed)
- `docs/adr/ADR-3-metaprompt-kernelization.md` — Metaprompt Kernelization
- `FLOSSI0ULLK_Master_Metaprompt_v1_3_1_Kernel.md` — Canonical kernel
- `docs/specs/harvest-protocol.spec.md` — HARVEST Protocol specification (minimal self-observation loop)
## Related Documents
- `docs/adr/ADR-0-recognition-protocol.md` — Recognition Protocol (Validated, all 4 criteria passed)
- `docs/adr/ADR-3-metaprompt-kernelization.md` — Metaprompt Kernelization
- `FLOSSI0ULLK_Master_Metaprompt_v1_3_1_Kernel.md` — Canonical kernel
- `docs/specs/harvest-protocol.spec.md` — HARVEST Protocol specification (minimal self-observation loop)
- `docs/governance/HARVEST_LOG.md` — HARVEST cycle evidence and validation outputs
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/adr/ADR-5-cognitive-virology-pattern.md` around lines 90 - 96, Add the
missing HARVEST log document to the ADR's Related Documents: in the Related
Documents section of ADR-5-cognitive-virology-pattern.md, append
`docs/governance/HARVEST_LOG.md — HARVEST log (validation trace/gate artifact)`
so the gate artifact referenced on line 70 is included for traceability; ensure
the entry mirrors the existing bullet style and provides a short descriptor for
verification.

kalisam and others added 3 commits March 25, 2026 09:33
Introduce a .serena directory with Serena configuration and project memories to onboard the assistant and document the codebase. Adds project.yml (project settings, language-server config, tool lists and defaults), several memory files (project_overview, codebase_structure, style_and_conventions, suggested_commands, task_completion_checklist) for developer guidance, and a .serena/.gitignore to ignore cache and project.local.yml.
New specs:
- phase0-substrate-bridge.spec.md: Formal SDD feature spec for the
  narrowest end-to-end validation test (publish, provenance, verify,
  query, fork-visible, no privilege). 6 testable criteria, 2-agent
  Tryorama implementation.

New research:
- Orchestration landscape v2.0.0 (summary + full report + Perplexity
  source). Supersedes v1.0.0 with empirical data: Silo-Bench k=50
  scalability wall, MAS-ProVe process verification failure, token
  economics collapse, IPFS availability decline.
- Cross-AI synthesis working doc moved from root to docs/research/.

Updated:
- MVP_PLAN.md: Added Architectural Constraints section with hard data
  (k<=5 agents, output verification > process, non-token economics,
  IPFS multi-pinning). Updated evidence sources and date. Immediate
  next action now points to substrate bridge spec.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Implements docs/specs/phase0-substrate-bridge.spec.md as a 2-agent
Tryorama test suite:

1. PUBLISH — assert_triple returns ActionHash
2. PROVENANCE — get_triple_record returns author pubkey + timestamp
3. VERIFY — Agent B retrieves Agent A's entry, confirms content + provenance
4. QUERY — Agent B discovers entry via query_triples without knowing hash
5. FORK-VISIBLE — Conflicting triples from two agents both persist
6. NO PRIVILEGE — Both agents use identical calls, can read each other

Also adds get_triple_record extern to coordinator (thin wrapper around
get() that exposes the full Record with signed_action for provenance
verification from the test side).

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
Copy link
Copy Markdown

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Warning

CodeRabbit couldn't request changes on this pull request because it doesn't have sufficient GitHub permissions.

Please grant CodeRabbit Pull requests: Read and write permission and re-run the review.

👉 Steps to fix this

Actionable comments posted: 18

Caution

Some comments are outside the diff and can’t be posted inline due to platform limitations.

⚠️ Outside diff range comments (2)
MVP_PLAN.md (2)

17-37: ⚠️ Potential issue | 🟡 Minor

Spec/entity counts are inconsistent within the same plan.

Line 17 states “6 entry types, 5 JSON schemas,” while Line 36 still says “5 entry type specs + 4 JSON schemas.” Please normalize these counts.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@MVP_PLAN.md` around lines 17 - 37, Update MVP_PLAN.md to use consistent
entity counts: find the occurrences of "docs/specs/ (6 entry types, 5 JSON
schemas, phase0 bridge spec)" and the later table row "5 entry type specs + 4
JSON schemas" and reconcile them to the correct authoritative numbers (either 6
entry types & 5 JSON schemas or 5 entry types & 4 JSON schemas). Replace both
places so they match exactly and ensure any related parenthetical text (e.g.,
"phase0 bridge spec") is preserved or updated accordingly; verify the numbers
against the actual contents of docs/specs/ before committing.

381-391: ⚠️ Potential issue | 🟠 Major

Immediate action and provided command are misaligned.

The text says to run Phase 0 bridge validation via a 2-agent Tryorama test, but the command block only performs cargo build. Add the actual test command(s) needed for the stated validation objective.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@MVP_PLAN.md` around lines 381 - 391, The markdown instructs running a 2-agent
Tryorama Phase 0 Substrate Bridge validation but the code block only shows a
cargo build; update the docs/specs/phase0-substrate-bridge.spec.md entry for
"Phase 0 Substrate Bridge Validation" to include the actual Tryorama test
invocation and prerequisite commands (e.g., nix develop, build wasm, start
conductor(s)/network if required) and the specific 2-agent Tryorama run command
that executes the test suite (referencing the Tryorama test file name or script,
e.g., the 2-agent test harness such as "tryorama run <path-to-phase0-test.js>"
or the npm/yarn script that triggers it), ensuring the single command sequence
in the block performs build + test end-to-end for the Phase 0 validation.
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In @.serena/memories/codebase_structure.md:
- Around line 42-46: The documented hot-path entries are stale: update the ARF
zome extern set (replace the generic "5 zome extern functions" in
ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs with the current exported
function names used by the runtime — e.g., the actual extern fn identifiers in
that file and any moved/renamed helpers like those referenced from budget.rs and
vector_ops.rs) and correct the bridge module path (replace
packages/orchestrator/holochain_connector.py with the current Python↔Holochain
connector module location/name). Locate the symbols in coordinator/src/lib.rs,
coordinator/src/budget.rs, coordinator/src/vector_ops.rs and the actual
connector module in packages/orchestrator to mirror current filenames and
exported function names in the markdown.

In @.serena/memories/project_overview.md:
- Around line 16-18: Update the architecture lines to match the actual code:
replace the Coordinator Zome extern list with the real exported functions (e.g.,
assert_triple and query_triples) referenced in the coordinator zome (search for
functions named assert_triple and query_triples in coordinator src/lib.rs) and
change the Python Orchestrator location to point to the actual connector module
(holochain_connector in the orchestrator package). Ensure the Coordinator entry
lists the correct extern function names and the Orchestrator entry references
holochain_connector so the doc matches the code.

In @.serena/memories/style_and_conventions.md:
- Around line 3-29: The MD022 lint warning is caused by missing blank lines
after ATX headings like "## Rust (Holochain Zomes)", "## Python", "## TypeScript
(Tests)", "## JSON Schema" and "## General"; fix by inserting a single blank
line immediately after each heading line so every heading is followed by an
empty line (e.g., add an empty line after the "## Rust (Holochain Zomes)"
heading and the same for "## Python", "## TypeScript (Tests)", "## JSON Schema"
and "## General") to satisfy markdownlint MD022.

In @.serena/memories/suggested_commands.md:
- Around line 3-44: Add missing blank lines around headings and fenced code
blocks to satisfy MD022/MD031: ensure there is a blank line before and after
each heading such as "## System (Windows + Git Bash)", "## Python", "## Rust /
Holochain (requires WSL2 + Nix)", "## Tryorama Integration Tests (inside WSL2
Nix shell)", "## Git", and "## Formatting/Linting", and ensure every
triple-backtick fenced block (e.g., ```bash blocks shown in the diff) has a
blank line immediately before the opening fence and immediately after the
closing fence so linting no longer reports heading/fence spacing issues.

In @.serena/memories/task_completion_checklist.md:
- Line 7: The checklist line "Update ADR if architectural decision was made" is
underspecified—replace or expand that item so it mandates the ADR include the
required sections: Problem Statement, Decision, Implementation Strategy,
Consequences, Validation Criteria, and Related Documents; update the checklist
entry (the list item text) to enumerate these fields explicitly (e.g., as
sub-checkboxes or a parenthetical) so reviewers must confirm each section is
present when closing a task.

In @.serena/project.yml:
- Around line 28-29: Update the project language list so editor/agent language
servers include Markdown and YAML support: modify the languages key (currently
only "python") to add "markdown" and "yaml" entries alongside "python" so LS
startup covers ADRs and spec files; ensure the languages array contains
"python", "markdown", and "yaml".
- Line 128: initial_prompt is empty and should contain the project
ADR/review-gate checklist to enforce the documented process; update the
initial_prompt key in .serena/project.yml to a non-empty YAML string that lists
the required gate items (e.g., ADR created/updated with link, ticket/issue
referenced, unit/integration tests added, CI passing, security/secret checks,
changelog entry, at least one reviewer/approval) and any commands or links used
by Serena; make sure the value is a properly quoted/multiline YAML string so it
remains intact and human-readable (reference: initial_prompt).

In `@docs/research/Automated-Agent-Orchestration-Report_v2.0.0-full.md`:
- Line 143: The phrase "4MB entry size limit" in the document should be
formatted with a space between the number and unit for consistency; update the
text fragment "4MB entry size limit" to "4 MB entry size limit" in the markdown
(look for the exact string "4MB entry size limit" in the docs/research content).
- Around line 29-33: The front-matter capability_truth_model is defined but not
applied consistently; update the document by tagging each major claim/component
with one of the four labels (verified, specified, aspirational, unverified) from
capability_truth_model — specifically add labels to items in "Key New Patterns"
(AMRO-S, CodeCRDT, AgentNet++, DecentLLMs), to entries in "Supporting
Infrastructure" (the systems currently marked "Production"), and ensure "Part 5:
Unsolved Problems" continues to show Claim Truth Model: Unverified; follow the
same format and verbatim labels used in the gap matrix examples so every claim
has an explicit truth status annotation for provenance and consistency.
- Around line 282-292: The Phase 0 Gate summary misses the "Query" criterion:
update the Phase 0 Gate text (the section titled "Phase 0 Gate (Unchanged,
Reinforced)") to explicitly include that agents must be able to discover
published provenance entries via querying (e.g., Agent B using query_triples
without prior knowledge of the artifact hash); add this as either a fourth
bullet in the 3-step summary or as an additional pass criterion alongside
"convergence across ≥3 nodes" so the report references the query_triples
discovery requirement by name.

In `@docs/research/Automated-Agent-Orchestration-Report_v2.0.0.md`:
- Around line 57-69: The "Impact" column in the updated comparison table lacks
inline evidence; update the Impact cells for entries such as "Agent scaling",
"Process verification", "IPFS availability", and "Token economics" to include
brief source references (e.g., "Design constraint: k<=5 (Silo-Bench §2.1)",
"Output verification > process (MAS-ProVe §3)", "Multi-pinning required (IPFS
Monitoring Report §4)", "All tokens declined (Market Data Q4 20XX)") so each
claim points to its supporting section or source; ensure the references follow
the same short-parenthetical style throughout the table and add a short
legend/note below the table explaining the citation format if needed.
- Around line 101-112: Update the integration status table to reflect that KERI
and hREA have existing bridge code (change their FLOSSI0ULLK Integration Status
from "Aspirational" to "Specified (code exists, not production-validated in
FLOSSI0ULLK context)"); reference the implemented artifacts
(AutonomousIdentifier, KeyEventLog, IdentitySeal entry types and functions
register_aid(), create_identity_seal(), rotate_key() in the
ARF/dnas/rose_forest/zomes/identity_integrity area) so the label clarifies
"Specified" (code present) while still noting not validated in FLOSSI0ULLK
production.

In `@docs/research/cross-ai-orchestration-synthesis-2026-03-25.md`:
- Around line 226-228: Extract the "gap matrix" block from the document titled
cross-ai-orchestration-synthesis-2026-03-25 (the well-structured mapping
currently labeled "gap matrix" that compares the v2.0.0 orchestration report
against FLOSSI0ULLK internal documents) into a new standalone markdown artifact;
add a metadata header including version, date, and status, explicitly cite the
source report (v2.0.0) and the internal documents reviewed, and include a short
"usage guidance" section describing how to consume the matrix for ADRs and
approvals; finally, replace the original inline matrix with a clear link/note
pointing to the new artifact and update any references in the synthesis section
that mention the matrix so they reference the new file.
- Around line 1-10: This document lacks a proper front-matter metadata header
and mixes working notes with formal analysis; add a structured metadata block at
the top (fields: version, status, date, truth_status, doc_type) and a clear
title indicating whether this is "Working Notes" or a "Formal Synthesis" (e.g.,
"# Cross-AI Orchestration Synthesis - Working Notes"); if you choose formal
synthesis, remove conversational artifacts such as the "Intent Echo:" paragraph
and checklist-style prompts and the conversational lines referenced in the
review (including the sections around the noted conversational markers and
prompts), and normalize tone/formatting to match other v2.0.0 reports;
alternatively, if keeping as working notes, explicitly mark doc_type as process
artifact and relocate or extract the gap matrix (the block around lines 564-674)
into its own properly structured document with title and metadata, leaving this
file as an index linking to that artifact.

In `@docs/research/Perplexity-Source-Agent-Orchestration-March2026.md`:
- Around line 5-454: The document contains widespread mojibake (e.g. sequences
like "—", "–", "â‰", "O(n²)") due to wrong/declared encoding; convert the
file to UTF-8 (ensure correct byte decoding), normalize Unicode (NFC), and run a
targeted find-and-replace for common corruptions replacing "—"→"—" (em dash),
"–"→"–" (en dash), "â‰"→"≠" or "≈" as context requires, and "O(n²)"→"O(n²)"
(and similar superscript/degree issues); verify math/inline formulas (e.g. the
pheromone update rule and tables) preserve intended symbols and run a
diff/spot-check on headings like the document title
"Perplexity-Source-Agent-Orchestration-March2026.md" and occurrences of "CRDTs",
"Holochain", "Silo-Bench" to confirm no semantic corruption.

In `@docs/specs/phase0-substrate-bridge.spec.md`:
- Around line 55-68: The fenced pseudocode blocks in the spec (containing calls
like assert_triple({...}), ASSERT lines, references to ActionHash and
get(action_hash)) should include a language specifier to improve readability and
editor support; update each triple/assertion pseudocode block to start with a
fenced code block marker that names a language (e.g., ```typescript or
```pseudo) so IDEs/highlighters treat the content appropriately, leaving the
pseudocode unchanged otherwise.
- Around line 162-168: Add a new Definition of Done checklist item that requires
implementation of provenance retrieval before running the substrate bridge
tests: require a provenance retrieval method (either the
get_triple_with_provenance extern or an extension to TripleResult) to be
implemented and validated so that tests referenced by substrate_bridge.test.ts
(notably test criteria 2, 3, and 5) can pass; ensure this DoD item is placed
before the existing "All 6 tests pass against running Holochain conductor" entry
and mention that its completion is prerequisite evidence for ADR-2 acceptance
and HARVEST_LOG.md recording.
- Around line 140-142: Choose option (b): extend the TripleResult struct to
include provenance fields (author/source and timestamp/created_at) instead of
introducing a separate get_triple_with_provenance extern; update the spec text
around query_triples to state that Query results include provenance, add
author/timestamp fields to TripleResult, and ensure the mapping from existing
KnowledgeTriple.source and KnowledgeTriple.created_at to the new TripleResult
fields is documented so implementers can return provenance directly from
query_triples.

---

Outside diff comments:
In `@MVP_PLAN.md`:
- Around line 17-37: Update MVP_PLAN.md to use consistent entity counts: find
the occurrences of "docs/specs/ (6 entry types, 5 JSON schemas, phase0 bridge
spec)" and the later table row "5 entry type specs + 4 JSON schemas" and
reconcile them to the correct authoritative numbers (either 6 entry types & 5
JSON schemas or 5 entry types & 4 JSON schemas). Replace both places so they
match exactly and ensure any related parenthetical text (e.g., "phase0 bridge
spec") is preserved or updated accordingly; verify the numbers against the
actual contents of docs/specs/ before committing.
- Around line 381-391: The markdown instructs running a 2-agent Tryorama Phase 0
Substrate Bridge validation but the code block only shows a cargo build; update
the docs/specs/phase0-substrate-bridge.spec.md entry for "Phase 0 Substrate
Bridge Validation" to include the actual Tryorama test invocation and
prerequisite commands (e.g., nix develop, build wasm, start conductor(s)/network
if required) and the specific 2-agent Tryorama run command that executes the
test suite (referencing the Tryorama test file name or script, e.g., the 2-agent
test harness such as "tryorama run <path-to-phase0-test.js>" or the npm/yarn
script that triggers it), ensuring the single command sequence in the block
performs build + test end-to-end for the Phase 0 validation.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: ASSERTIVE

Plan: Pro

Run ID: bf804a07-3329-46c1-89df-1620db152a72

📥 Commits

Reviewing files that changed from the base of the PR and between 10ebfce and 36ea046.

📒 Files selected for processing (13)
  • .serena/.gitignore
  • .serena/memories/codebase_structure.md
  • .serena/memories/project_overview.md
  • .serena/memories/style_and_conventions.md
  • .serena/memories/suggested_commands.md
  • .serena/memories/task_completion_checklist.md
  • .serena/project.yml
  • MVP_PLAN.md
  • docs/research/Automated-Agent-Orchestration-Report_v2.0.0-full.md
  • docs/research/Automated-Agent-Orchestration-Report_v2.0.0.md
  • docs/research/Perplexity-Source-Agent-Orchestration-March2026.md
  • docs/research/cross-ai-orchestration-synthesis-2026-03-25.md
  • docs/specs/phase0-substrate-bridge.spec.md

Comment on lines +42 to +46
- `ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs` — 5 zome extern functions
- `ARF/dnas/rose_forest/zomes/coordinator/src/budget.rs` — Source-chain budget tracking
- `ARF/dnas/rose_forest/zomes/coordinator/src/vector_ops.rs` — Cosine similarity math
- `ARF/dnas/rose_forest/zomes/integrity/src/lib.rs` — Entry types + validation rules
- `packages/orchestrator/holochain_connector.py` — Python↔Holochain bridge
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Hot-path references should be corrected to current symbols/locations.

This section appears out of sync with implementation details (extern function set and connector module path), which weakens the usefulness of this map for onboarding/debugging.

🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 43-43: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 44-44: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.serena/memories/codebase_structure.md around lines 42 - 46, The documented
hot-path entries are stale: update the ARF zome extern set (replace the generic
"5 zome extern functions" in ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs
with the current exported function names used by the runtime — e.g., the actual
extern fn identifiers in that file and any moved/renamed helpers like those
referenced from budget.rs and vector_ops.rs) and correct the bridge module path
(replace packages/orchestrator/holochain_connector.py with the current
Python↔Holochain connector module location/name). Locate the symbols in
coordinator/src/lib.rs, coordinator/src/budget.rs, coordinator/src/vector_ops.rs
and the actual connector module in packages/orchestrator to mirror current
filenames and exported function names in the markdown.

Comment on lines +16 to +18
- **Coordinator Zome** (`ARF/dnas/rose_forest/zomes/coordinator/`): 5 extern functions (add_knowledge, vector_search, link_edge, budget_status, create_thought_credential)
- **Budget System**: Agent-local RU tracking via source chain queries, 100 RU/24h window
- **Python Orchestrator** (`packages/orchestrator/`): Async WebSocket connector to Holochain conductor
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Architecture overview is inconsistent with current code paths/APIs.

The coordinator extern list and orchestrator location here don’t match the provided code snippets (assert_triple/query_triples in ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs, and connector at ARF/in.finite-nrg/infinity-bridge/orchestrator/holochain_connector.py). Please align these lines to avoid misleading implementation docs.

🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 17-17: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 18-18: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.serena/memories/project_overview.md around lines 16 - 18, Update the
architecture lines to match the actual code: replace the Coordinator Zome extern
list with the real exported functions (e.g., assert_triple and query_triples)
referenced in the coordinator zome (search for functions named assert_triple and
query_triples in coordinator src/lib.rs) and change the Python Orchestrator
location to point to the actual connector module (holochain_connector in the
orchestrator package). Ensure the Coordinator entry lists the correct extern
function names and the Orchestrator entry references holochain_connector so the
doc matches the code.

Comment on lines +3 to +29
## Rust (Holochain Zomes)
- **Entry types**: Defined in integrity zome with `#[hdk_entry_helper]` + `#[derive(Clone, PartialEq)]`
- **Validation**: In integrity zome's `validate()` — returns `ValidateCallbackResult::Valid` or `Invalid(reason)`
- **Error codes**: Prefixed with `E_` (e.g., `E_BUDGET_EXCEEDED`, `E_LICENSE`, `E_EMBED_DIM`)
- **Link types**: Minimal — only `AllNodes` and `Edge` (removed ShardMember and AgentBudget as Holochain-native alternatives exist)
- **Budget state**: Queried from local source chain via `ChainQueryFilter`, NOT from DHT links
- **Constants**: SCREAMING_SNAKE_CASE with doc comments explaining the "why"
- **Paths**: Used for DHT discovery (e.g., `Path::from("all_nodes")`)

## Python
- **Type hints**: Expected but not enforced everywhere yet
- **Async**: Used in orchestrator (asyncio + websockets)
- **Testing**: pytest, with offline and integration test separation
- **Serialization**: msgpack for Holochain wire protocol
- **Naming**: snake_case for functions/variables, PascalCase for classes

## TypeScript (Tests)
- **Framework**: Vitest + Tryorama 0.17
- **AJV import**: `import Ajv from "ajv/dist/2020.js"` (NOT `"ajv"` — draft 2020-12)
- **libsodium fix**: Postinstall symlink script (`fix-libsodium-esm.mjs`)

## JSON Schema
- Draft 2020-12 (`$schema: "https://json-schema.org/draft/2020-12/schema"`)
- Located in `docs/specs/`
- Spec-code linkage: RoseNode's `model_card_hash` = SHA-256 of spec prose

## General
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Address heading spacing warnings to keep docs lint green.

Add blank lines after headings per MD022 so this conventions doc passes markdownlint consistently.

🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 3-3: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 12-12: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 19-19: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 24-24: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 29-29: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.serena/memories/style_and_conventions.md around lines 3 - 29, The MD022
lint warning is caused by missing blank lines after ATX headings like "## Rust
(Holochain Zomes)", "## Python", "## TypeScript (Tests)", "## JSON Schema" and
"## General"; fix by inserting a single blank line immediately after each
heading line so every heading is followed by an empty line (e.g., add an empty
line after the "## Rust (Holochain Zomes)" heading and the same for "## Python",
"## TypeScript (Tests)", "## JSON Schema" and "## General") to satisfy
markdownlint MD022.

Comment on lines +3 to +44
## System (Windows + Git Bash)
- `ls`, `cd`, `grep`, `find` — standard Unix commands via Git Bash
- `git` — version control
- `curl` — HTTP requests (no `gh` CLI available)

## Python
```bash
cd ARF && pip install -e .
pytest # All Python tests
pytest tests/test_committee_validation.py -v # Specific test
pytest -k "memory" # Pattern match
python -m pytest --cov=. --cov-report=html # Coverage
```

## Rust / Holochain (requires WSL2 + Nix)
```bash
# Enter Nix dev shell (inside WSL2)
cd ARF && nix develop
cargo test --lib # Rust unit tests
cargo build --release --target wasm32-unknown-unknown # WASM build
hc dna pack workdir/dna # Pack DNA
hc app pack workdir/ # Pack hApp
```

## Tryorama Integration Tests (inside WSL2 Nix shell)
```bash
cd ARF/tests/tryorama
npm install # Installs deps + runs postinstall fix
npx vitest run # Run all integration tests
npx vitest run spec_compliance.test.ts # Specific test file
```

## Git
```bash
git status # Check state
git log --oneline -20 # Recent history
git diff # Unstaged changes
git fetch origin # Fetch remote
```

## Formatting/Linting
```bash
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟡 Minor

Fix markdownlint heading/fence spacing warnings.

Headings and fenced code blocks need surrounding blank lines (MD022/MD031), otherwise docs lint will keep failing.

Suggested markdown spacing fix
 # Suggested Commands
 
 ## System (Windows + Git Bash)
+
 - `ls`, `cd`, `grep`, `find` — standard Unix commands via Git Bash
 - `git` — version control
 - `curl` — HTTP requests (no `gh` CLI available)
 
 ## Python
+
 ```bash
 cd ARF && pip install -e .
 pytest                                    # All Python tests
@@
 python -m pytest --cov=. --cov-report=html  # Coverage

Rust / Holochain (requires WSL2 + Nix)

# Enter Nix dev shell (inside WSL2)
@@
hc app pack workdir/                     # Pack hApp

Tryorama Integration Tests (inside WSL2 Nix shell)

@@
npx vitest run spec_compliance.test.ts   # Specific test file

Git

@@
git fetch origin                         # Fetch remote

Formatting/Linting

# Python
🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 3-3: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 8-8: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 9-9: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)


[warning] 17-17: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 18-18: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)


[warning] 27-27: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 28-28: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)


[warning] 35-35: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 36-36: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)


[warning] 43-43: Headings should be surrounded by blank lines
Expected: 1; Actual: 0; Below

(MD022, blanks-around-headings)


[warning] 44-44: Fenced code blocks should be surrounded by blank lines

(MD031, blanks-around-fences)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.serena/memories/suggested_commands.md around lines 3 - 44, Add missing
blank lines around headings and fenced code blocks to satisfy MD022/MD031:
ensure there is a blank line before and after each heading such as "## System
(Windows + Git Bash)", "## Python", "## Rust / Holochain (requires WSL2 + Nix)",
"## Tryorama Integration Tests (inside WSL2 Nix shell)", "## Git", and "##
Formatting/Linting", and ensure every triple-backtick fenced block (e.g.,
```bash blocks shown in the diff) has a blank line immediately before the
opening fence and immediately after the closing fence so linting no longer
reports heading/fence spacing issues.

2. **Check for lint/format issues** — black + ruff (Python), cargo fmt + clippy (Rust)
3. **Review against spec** — ensure implementation matches JSON Schema specs in `docs/specs/`
4. **Check for removed wheel-reinvention** — don't duplicate what Holochain provides natively (DHT distribution, source chain versioning, validation callbacks)
5. **Update ADR if architectural decision was made**
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🛠️ Refactor suggestion | 🟠 Major

Make the ADR checklist item explicit about required sections.

“Update ADR” is underspecified; add the mandatory ADR fields so this checklist enforces completeness, not just existence.

Proposed checklist wording
-5. **Update ADR if architectural decision was made**
+5. **Update/Create ADR for major decisions** — include: Problem Statement, Decision, Implementation Strategy, Consequences (Positive/Negative/Neutral), Validation Criteria, Related Documents

Based on learnings: ADRs for major decisions must include Problem Statement, Decision, Implementation Strategy, Consequences, Validation Criteria, and Related Documents.

📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
5. **Update ADR if architectural decision was made**
5. **Update/Create ADR for major decisions** — include: Problem Statement, Decision, Implementation Strategy, Consequences (Positive/Negative/Neutral), Validation Criteria, Related Documents
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.serena/memories/task_completion_checklist.md at line 7, The checklist line
"Update ADR if architectural decision was made" is underspecified—replace or
expand that item so it mandates the ADR include the required sections: Problem
Statement, Decision, Implementation Strategy, Consequences, Validation Criteria,
and Related Documents; update the checklist entry (the list item text) to
enumerate these fields explicitly (e.g., as sub-checkboxes or a parenthetical)
so reviewers must confirm each section is present when closing a task.

Comment on lines +226 to +228

If you want, I’ll turn this into a strict **gap matrix**: `report claim -> matching internal doc -> contradiction -> promote/hold/reject`.

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🧹 Nitpick | 🔵 Trivial

Extract gap matrix as separate formal artifact.

Lines 564-674 contain a well-structured gap matrix mapping the v2.0.0 orchestration report against FLOSSI0ULLK internal documents with Promote/Hold/Reject classifications. This is the concrete "bridge artifact between landscape awareness and build authorization" referenced in the synthesis (line 268).

Recommendation: Extract the gap matrix to a separate file (e.g., docs/governance/orchestration-v2.0.0-gap-matrix.md) with:

  • Proper metadata header (version, date, status)
  • Reference to source report (v2.0.0) and internal docs checked
  • Clear usage guidance

This would make the gap matrix easier to reference in ADRs and align with the project's document organization patterns.

🧰 Tools
🪛 LanguageTool

[style] ~226-~226: This phrasing can be overused. Try elevating your writing with a more formal alternative.
Context: ...DR-1’s premature-integration mistake. If you want, I’ll turn this into a strict **gap mat...

(IF_YOU_WANT)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/research/cross-ai-orchestration-synthesis-2026-03-25.md` around lines
226 - 228, Extract the "gap matrix" block from the document titled
cross-ai-orchestration-synthesis-2026-03-25 (the well-structured mapping
currently labeled "gap matrix" that compares the v2.0.0 orchestration report
against FLOSSI0ULLK internal documents) into a new standalone markdown artifact;
add a metadata header including version, date, and status, explicitly cite the
source report (v2.0.0) and the internal documents reviewed, and include a short
"usage guidance" section describing how to consume the matrix for ADRs and
approvals; finally, replace the original inline matrix with a clear link/note
pointing to the new artifact and update any references in the synthesis section
that mention the matrix so they reference the new file.

Comment on lines +5 to +454
The convergence of multi-agent AI systems with decentralized infrastructure presents both a compelling architectural vision and a sobering set of unsolved engineering problems. This report synthesizes evidence from academic research, production deployments, and open-source projects to evaluate the current state of automated agent orchestration for decentralized AI development. The findings are mixed: coordination algorithms like Contract Net Protocol, ant colony optimization, and CRDTs have demonstrated real performance gains in controlled settings — [CodeCRDT achieved 100% convergence with zero merge failures](https://arxiv.org/pdf/2510.18893) across 600 trials, and [AMRO-S delivered 4.7x speedup](https://arxiv.org/html/2603.12933) at 1,000 concurrent agents — but production deployments remain concentrated in narrow domains. [Autonolas has processed 8.8 million agent-to-agent transactions](https://olas.network/timeline), making it the most credible implementation, yet nearly all that activity is in DeFi prediction markets rather than broader automation use cases.

The infrastructure layer shows genuine maturity in specific areas: [Holochain's 0.6.0 release](https://blog.holochain.org/2025-at-a-glance-landing-reliability/) delivered Warrants for Byzantine agent isolation without global consensus, [AD4M's v0.10.1](https://github.com/coasys/ad4m/releases) integrated local AI inference into its agent-centric meta-ontology, and [IPFS's DASL initiative](https://ipfsfoundation.org/content-addressing-2025-in-review/) improved cross-implementation interoperability. However, the critical finding from [Silo-Bench (March 2026)](https://arxiv.org/html/2603.01045v1) is that at k=50 agents, coordination overhead eliminates all parallelization gains entirely. This scalability wall defines the frontier: decentralized agent orchestration must solve coordination overhead at scale before the vision of autonomous agent economies becomes viable. The report recommends a layered architecture — Holochain for trust, CRDTs for state synchronization, gossip protocols for discovery, contract net for task allocation, and reputation-weighted governance — as the most evidence-supported path forward, while being explicit about the substantial gaps that remain.

## Multi-Agent Coordination Algorithms and Their Trade-Offs

### Contract Net Protocol and Auction-Based Allocation

The Contract Net Protocol (CNP), [introduced by Reid G. Smith in 1980](https://en.wikipedia.org/wiki/Contract_Net_Protocol), remains the most widely deployed task allocation mechanism for heterogeneous multi-agent systems. It operates through a structured negotiation cycle: a Manager agent broadcasts a call for proposals, Contractor agents submit bids reflecting their capability and availability, and the Manager awards the contract to the best bidder. The protocol enables recursive decomposition — a Contractor can itself become a Manager for subtasks, creating [self-organizing hierarchies without rigid top-down control](https://notes.muthu.co/2025/10/task-allocation-for-machine-teamwork-with-the-contract-net-protocol/).

Auction variants formalize CNP with game-theoretic guarantees. [Vickrey (second-price) auctions achieve incentive compatibility](https://notes.muthu.co/2025/10/auction-based-task-allocation-in-multi-agent-systems/) — the dominant strategy is truthful bidding — while combinatorial auctions allow agents to bid on task bundles for synergistic allocations, though optimal clearing is NP-hard. The [Greedy Coalition Auction Algorithm (GCAA)](https://arxiv.org/abs/2107.00144) achieves convergence after at most N iterations for N agents.

CNP maps directly onto modern multi-agent frameworks: the orchestrator-worker pattern in [LangGraph and CrewAI](https://sam-solutions.com/blog/multi-agent-orchestration/) is a direct analogue. For open-source AI agent orchestration, CNP is well-suited when agents are heterogeneous, task arrival is dynamic, and privacy constraints prevent sharing full agent state. Its weakness is communication overhead from broadcasting and vulnerability to strategic manipulation by dishonest bidders without enforcement mechanisms.

### BDI Architecture and Cognitive Agent Models

The Belief-Desire-Intention (BDI) architecture, formalized from [Michael Bratman's theory of practical reasoning](https://en.wikipedia.org/wiki/Belief%E2%80%93desire%E2%80%93intention_software_model), provides the most explainable coordination framework. BDI agents maintain three cognitive state components — beliefs (world model), desires (objectives), and intentions (committed plans) — cycling through perception, deliberation, and execution. The separation of plan selection from plan execution enables agents to [balance deliberation time against reactive responsiveness](https://smythos.com/developers/agent-development/agent-oriented-programming-and-bdi-agents/).

Modern extensions integrate BDI with probabilistic reasoning (POMDP-BDI hybridization), reinforcement learning ([AMAD-SRL for drone agents](https://www.emergentmind.com/topics/bdi-architectures)), and classical planning to [dynamically generate new plans](https://www.meneguzzi.eu/felipe/pubs/ecai-bdi-plan-2024.pdf) when pre-compiled plan libraries are insufficient. For multi-agent coordination, BDI agents can coordinate through [social practices structured as partially ordered landmarks](https://www.emergentmind.com/topics/bdi-architectures) and through [Theory of Mind — recognizing other agents' intentions from observed behavior](https://www.ifaamas.org/Proceedings/aamas2024/pdfs/p2679.pdf).

BDI maps naturally onto LLM-based agent architectures: the plan library corresponds to retrieved tool chains, intentions to committed multi-step workflows, and beliefs to context/memory. The [2026 BDI-cooperative agents paper](https://www.scitepress.org/Papers/2026/142839/142839.pdf) demonstrates successful BDI deployment in resource-constrained cooperative MAS. BDI is highly suitable when explainability and auditability are requirements — critical for open-source governance where agent behavior must be inspectable.

### Swarm Intelligence Applied to LLM Agent Routing

Swarm intelligence algorithms offer the most scalable coordination mechanisms by leveraging indirect, environment-mediated communication. Ant Colony Optimization (ACO), [formalized by Marco Dorigo in 1992](https://www.distributedthoughts.org/digital-pheromones-what-ants-know-about-agent-coordination/), uses pheromone-mediated routing where solution quality reinforces path selection while evaporation prevents stagnation. The pheromone update rule — \\(\tau_{ij}(t+1) = (1 - \rho) \cdot \tau_{ij}(t) + \Delta\tau_{ij}\\) — enables decentralized optimization without centralized controllers.

[AMRO-S (March 2026)](https://arxiv.org/html/2603.12933) applies ACO directly to LLM multi-agent routing, modeling the agent pool as a layered directed graph where each node represents a backbone model paired with a reasoning policy. Using task-specific pheromone matrices, quality-gated asynchronous updates, and a semantic small language model router, AMRO-S achieves up to **4.7x speedup** under 1,000 concurrent processes compared to strong routing baselines.

Particle Swarm Optimization (PSO) offers a complementary approach for continuous optimization. [LMPSO (April 2025)](https://arxiv.org/pdf/2504.09247) treats each particle's velocity as a prompt that generates the next candidate solution via an LLM, outperforming PSO variants on problems requiring flexible, context-rich representations. Both algorithms are suited to massive-scale routing and load balancing. The key limitation is convergence — neither guarantees finding the global optimum, and [stigmergic communication requires a shared substrate](https://www.distributedthoughts.org/digital-pheromones-what-ants-know-about-agent-coordination/) with its own consistency challenges.

### Multi-Agent Reinforcement Learning (MARL): CTDE and Beyond

MARL applies reinforcement learning to systems of interacting agents. The dominant paradigm is [Centralized Training for Decentralized Execution (CTDE)](https://arxiv.org/abs/2409.03052): training uses global state information while execution uses only local observations. Key algorithms include QMIX (factored joint Q-function), MADDPG (centralized critics), and MAPPO (shared critic with PPO). The [CADP framework (IJCAI 2025)](https://www.ijcai.org/proceedings/2025/803) introduces "Centralized Advising and Decentralized Pruning" — agents exchange advice during training via explicit communication channels, then progressively eliminate this communication for deployment, achieving superior performance versus CTDE baselines.

[TAG (February 2025)](https://arxiv.org/abs/2502.15425) addresses hierarchical MARL in a fully decentralized setting, demonstrating that agents can learn coordinated hierarchical policies without any centralized training component. This is significant for open-source deployments where a centralized training phase may be infeasible.

The central limitation of CTDE for open-source agent orchestration is that agents trained together develop implicit coordination that breaks when any agent is replaced by a different version. In ecosystems where agent identity is fluid and versioning is continuous, MARL policies become fragile. Independent learners (DTE) avoid this but suffer from non-stationarity.

### Blackboard Systems and Stigmergic Coordination

Blackboard systems coordinate agents through shared state rather than direct communication. Agents read a shared data structure, contribute specialized reasoning, and the current blackboard state determines activation. [Google Research's Blackboard MAS (2025)](https://research.google/pubs/blackboard-multi-agent-systems-for-information-discovery-in-data-science/) achieves 13–57% improvement over RAG and master-slave baselines on data discovery benchmarks. The [LbMAS framework](https://arxiv.org/html/2507.01701v1) adds conflict-resolution agents that detect contradictions, trigger private discussion, and post reconciled outputs.

The weakness is logical centralization — the shared blackboard becomes a bottleneck at scale, requiring CRDTs or consensus machinery for distributed deployments.

Stigmergic coordination goes further by eliminating even the centralized shared state. Agents modify the environment, and those modifications influence subsequent agent behavior — exactly how [insect colonies coordinate through pheromone trails](https://www.sciencedirect.com/science/article/abs/pii/S0166361503001234). A [March 2026 analysis on LessWrong](https://www.lesswrong.com/posts/sX9LztxjtSEwd8qEo/emergent-stigmergic-coordination-in-ai-agents-1) documented emergent stigmergy in LLM web-search agents: each agent's search behavior modified web indices, and subsequent agents detected and adapted to those modifications — an unintended but revealing illustration of stigmergic coordination at web scale.

As [Distributed Thoughts (February 2026)](https://www.distributedthoughts.org/digital-pheromones-what-ants-know-about-agent-coordination/) observes: "every major framework — CrewAI, LangGraph, AutoGen — follows direct coordination, and every one hits the same scaling wall." Stigmergy eliminates the O(n²) communication overhead but introduces unpredictability — emergent dynamics are hard to reason about and convergence is not guaranteed.

### Market-Based Mechanisms and Mechanism Design

Market-based coordination treats agent systems as economies where price signals coordinate decentralized allocation. Vickrey-Clarke-Groves (VCG) mechanisms achieve social welfare maximization through truthful bidding, while [double auctions enable continuous markets](https://sam-solutions.com/blog/multi-agent-orchestration/) where buyer and seller agents submit bids/asks simultaneously. For open-source AI ecosystems where agents have heterogeneous capability costs and resource constraints, market mechanisms with subsidized prices or [quadratic funding](https://gitcoin.co/mechanisms/conviction-voting) for public goods are natural fits.

[Autonolas's Mech Marketplace](https://olas.network/timeline) represents the most mature production implementation: an on-chain agent-to-agent service marketplace where agents hire, sell services, and collaborate autonomously, with over 8.8 million transactions processed. The marketplace uses [Tendermint-like consensus](https://olas.network/blog/olas-the-2021-origin-story-of-one-of-the-first-crypto-ai-agent-projects-now-scaling-past-9-9-million-a2-a-transactions) for off-chain agent coordination combined with on-chain settlement.

### Comparative Analysis Table

| Algorithm | Decentralization Level | Fault Tolerance | Scalability | Latency | Comm. Overhead | Suitability for Open-Source AI |
|---|---|---|---|---|---|---|
| **Contract Net / Auction** | Moderate (Manager per task) | High (self-selection) | High — O(n) per task | Broadcast + bid wait | O(n) per announcement | High: dynamic, heterogeneous agents |
| **BDI Architecture** | Low–Moderate | High (intention revision) | Moderate (< 100 agents) | Low (local reasoning) | Low (message-passing) | High for explainability; limited scale |
| **ACO / Swarm** | Very High | Very High | Very High (1000+) | Variable (convergence) | Low (environment only) | Excellent for routing/load balancing |
| **MARL (CTDE)** | Execution only | High (decentralized exec) | High | Training: high; Exec: low | Zero at execution | Poor: agent replacement breaks coordination |
| **Blackboard** | Low (shared state) | Moderate | Limited by throughput | Per-round overhead | Low (indirect) | Good for complex problem-solving |
| **Stigmergy** | Very High | Very High | Very High | Asynchronous | Very Low | Excellent for massive async networks |
| **Market-Based** | High (price emergence) | Moderate | High | Market clearing time | Moderate | Good with payment infrastructure |
| **[AgentNet++ Hierarchical](https://arxiv.org/abs/2512.00614)** | High (hierarchical) | High | 1000+ agents | 40% less overhead | Low | 23% higher task completion |
| **[Gossip-Based](https://arxiv.org/abs/2512.03285)** | Very High | Very High | High | Eventual | O(n log n) | Excellent under uncertainty |

## Distributed Consensus and Governance Mechanisms

### Classical Consensus (Raft, Paxos) vs. Byzantine Tolerance (PBFT, HotStuff, Tendermint)

[Raft consensus](https://raft.github.io), designed for understandability, decomposes consensus into leader election, log replication, and safety guarantees. It tolerates f crashed nodes in a 2f+1 system — a 5-agent cluster survives 2 failures. [Raft has been applied directly to multi-agent formation control](https://arxiv.org/pdf/2308.10097), with each agent as a Raft node and position updates replicated as log entries. Production deployments include etcd (Kubernetes), CockroachDB, and TiKV.

[Paxos](https://www.sciencedirect.com/topics/computer-science/paxos-algorithm) provides stronger theoretical foundations but extreme implementation complexity. [Multi-Paxos reduces message overhead from 4 RTTs to 2 RTTs](https://arpitbhayani.me/blogs/multi-paxos/) in steady state. Both Raft and Paxos handle only crash failures — a single malicious agent can corrupt the entire system.

For open multi-agent networks with potentially adversarial agents, Byzantine Fault Tolerance (BFT) is essential. The fundamental theorem: n = 3f + 1 nodes tolerate f Byzantine nodes. [PBFT](https://decentralizedthoughts.github.io/2019-06-23-what-is-the-difference-between/) achieves this with O(n²) communication but limits practical deployments to ~20–50 nodes. [HotStuff](https://decentralizedthoughts.github.io/2019-06-23-what-is-the-difference-between/) achieves linear O(n) communication using threshold signatures, powering Aptos and Meta's LibraBFT. [Tendermint](https://infoscience.epfl.ch/bitstreams/bb494e9a-22aa-43a2-b995-69c7a2cc893e/download) optimizes for P2P gossip networks with zero-complexity view change and strong accountability through proof-of-fork — it powers the Cosmos ecosystem and [Autonolas's off-chain agent consensus](https://olas.network/timeline).

[DecentLLMs (July 2025)](https://arxiv.org/html/2507.14928v1) applies leaderless BFT to LLM multi-agent systems: worker agents generate answers in parallel while evaluator agents score using the Geometric Median algorithm, which is provably Byzantine-robust. This eliminates the leader-targeted attack vector present in PBFT.

The [Swarm Contract framework](https://arxiv.org/abs/2412.19256) proposes multi-sovereign agent consensus within Trusted Execution Environments (TEEs), enabling agents from different organizations to reach agreement with cryptographic attestation of correct execution — a promising pattern for cross-organizational agent orchestration.

### DAG-Based Consensus and Agent-Centric Validation

[Hashgraph](https://101blockchains.com/blockchain-vs-hashgraph-vs-dag-vs-holochain/) achieves asynchronous BFT consensus through "gossip-about-gossip" and virtual voting — nodes deterministically compute what consensus would be without sending actual votes. IOTA's Tangle uses a pay-it-forward model where each transaction validates two previous ones, enabling fee-free micro-transactions suitable for IoT-scale agent networks but with [weaker security thresholds (1/3 attack vs. 50%)](https://tokens-economy.gitbook.io/consensus/chain-based-dag/direct-acyclic-graph-tangle-dag).

[AgentNet (NeurIPS 2025)](https://neurips.cc/virtual/2025/poster/115584) uses a DAG-structured network where agents are nodes and edges represent task dependencies, achieving fault tolerance and scalable specialization. [AgentNet++ extends this with hierarchical decentralization](https://arxiv.org/abs/2512.00614), demonstrating 23% higher task completion rates and 40% less communication overhead while scaling to 1,000+ agents.

### Holochain's Agent-Centric Model: Scaling Without Global Consensus

Holochain represents a fundamental architectural inversion: rather than a single global state requiring consensus before acceptance, [each agent maintains their own source chain](https://www.holochain.org/documents/holochain-white-paper-2.0.pdf) — a personal append-only log — and shares data through a sharded DHT. Validation is peer-distributed using DNA-encoded rules rather than network-wide.

The [Holochain White Paper 2.0](https://www.holochain.org/documents/holochain-white-paper-2.0.pdf) formalizes this as "Scaling Consent" — coherent collaborative action without global consensus. Performance benchmarks from [PMC/Sensors (June 2025)](https://pmc.ncbi.nlm.nih.gov/articles/PMC12251913/) demonstrate 50ms publish latency (vs. 200ms for blockchain), 30ms retrieve latency (vs. 100ms), and throughput that scales horizontally as agents are added rather than degrading.

The critical limitation is that Holochain cannot natively solve double-spend problems or provide total ordering — applications requiring global agreement need additional mechanisms. As the [Ethereum Research discussion](https://ethresear.ch/t/holochain-an-agent-centric-framework-for-distributed-apps/5153) clarifies: "Holochain applications maintain systemic integrity without consensus... because any single node uses provenance to independently verify any single transaction."

### CRDTs: Coordination-Free Distributed State

Conflict-free Replicated Data Types (CRDTs), [formally defined by Shapiro et al. in 2011](https://en.wikipedia.org/wiki/Conflict-free_replicated_data_type), provide mathematically guaranteed convergence across replicas without coordination. The merge function satisfies commutativity, associativity, and idempotence, ensuring that any two replicas receiving the same set of updates converge to the same state regardless of order.

[CodeCRDT (October 2025)](https://arxiv.org/pdf/2510.18893) applies CRDTs to concurrent LLM agent code generation using a Yjs CRDT document. Multiple implementation agents work in parallel with a TODO-claim protocol — agents write ownership claims to a CRDT map with LWW semantics resolving concurrent claims deterministically. Results: **100% convergence, zero character-level merge failures** across 600 trials, with 21.1% speedup on some tasks and median propagation latency of 50ms.

The [10-year retrospective on coordination-free programming](https://dl.acm.org/doi/10.1145/3756907.3756910) documents CRDTs operating at 1,000+ AWS nodes for the Lasp system, demonstrating that coordination-free distributed state is viable at production scale. Production CRDT deployments include Redis, Riak, Azure CosmosDB, and Figma.

CRDTs are ideal for shared agent task queues, distributed configuration, collaborative generation, and voting/counting mechanisms. They are inappropriate for scenarios requiring strict linearizability — financial settlement and irreversible actions still require consensus mechanisms.

### Governance for Open-Source Projects

**Conviction Voting** is a continuous governance mechanism where voting power accumulates over time through sustained token staking. [Originally derived from Michael Zargham's research](https://gitcoin.co/mechanisms/conviction-voting) and implemented by Commons Stack and 1Hive, it rewards long-term commitment and makes whale attacks costly. [Polkadot OpenGov](https://gitcoin.co/mechanisms/conviction-voting) uses conviction-based time-lock multipliers (lock 28 days = 2x voting weight), producing an order-of-magnitude increase in votes per referendum.

**Quadratic Voting** expresses preference intensity while limiting plutocratic dominance — the cost of n votes is n², so each additional vote is progressively more expensive. The Sybil vulnerability (splitting wallets to reduce quadratic cost) requires identity verification; [ConVo (DAWO 2024)](https://dawo24.org/wp-content/uploads/2024/06/Abstract_32.pdf) proposes combining quadratic voting with Proof-of-Personhood biometrics.

**Token-Curated Registries (TCRs)** govern decentralized lists through staking and challenge mechanisms. [Multicoin Capital's analysis](https://multicoin.capital/2018/09/05/tcrs-features-and-tradeoffs/) identifies failure modes including coordination attacks and voter apathy, but TCRs remain well-suited for curating verified agent capability registries.

**Reputation-Based Governance** derives voting power from non-transferable reputation earned through contributions. [Colony.io's implementation](https://blog.colony.io/what-is-reputation-based-governance) features domain-specific reputation, temporal decay, and smart-contract-managed scores. Compared to [token-weighted systems](https://www.chainscorelabs.com/en/blog/network-states-and-pop-up-cities/network-state-governance-models/why-reputation-based-governance-will-outperform-token-weighted-voting), reputation-based governance provides stronger Sybil resistance, expertise alignment, and plutocracy resistance, though it faces a cold-start problem.

**Futarchy** separates values (chosen democratically) from policies (determined by prediction markets). [Meta-DAO on Solana](https://defiprime.com/futard-prediction-markets) is the first production futarchical system, and [Frontiers in Blockchain (October 2025)](https://www.frontiersin.org/journals/blockchain/articles/10.3389/fbloc.2025.1650188/full) demonstrated alignment between retrospective futarchic simulation and historical DeSci DAO decisions. The limitation is Goodhart's Law — poorly chosen welfare metrics can pervert outcomes.

[AgentDAO](https://arxiv.org/abs/2503.10099) directly applies LLM multi-agent systems to DAO governance: specialized agents analyze proposals, generate arguments, and vote. This represents an early experiment in using AI agents as governance participants rather than merely governed entities.

### Comparative Analysis Table

| Mechanism | Fault Model | Communication | Scalability | Decentralization | Best Use Case |
|---|---|---|---|---|---|
| **Raft** | Crash (f in 2f+1) | O(n) | Moderate | Leader-based | Agent cluster state replication |
| **PBFT** | Byzantine (f in 3f+1) | O(n²) | Low (20–50 nodes) | Leader-based | Small trusted agent committees |
| **HotStuff** | Byzantine | O(n) | High | Rotating leader | Large untrusted networks |
| **Tendermint** | Byzantine | O(n) | High | Rotating leader | Agent consensus with accountability |
| **Holochain** | Agent-centric | Local validation | Unlimited | Full | Agent-private data; no global ordering |
| **CRDTs** | Partition-tolerant | Zero coordination | Very High | Full | Concurrent shared state |
| **Conviction Voting** | N/A | Continuous staking | High | Token-weighted | Treasury allocation |
| **Quadratic Voting** | N/A | Budget allocation | High | Identity-required | Priority ranking |
| **Reputation (Colony)** | N/A | Domain-scoped | High | Earned/non-transferable | Expert technical decisions |
| **Futarchy** | N/A | Market trading | Moderate | Capital-weighted | High-stakes policy decisions |

## Infrastructure Patterns Enabling Value Alignment Without Centralization

### Holochain: Trust Foundation

Holochain's DNA validation model provides the most granular trust mechanism for decentralized agent orchestration. A Holochain application's DNA bundles WebAssembly code into [integrity zomes (deterministic validation rules) and coordinator zomes (business logic)](https://developer.holochain.org/concepts/7_validation/). Every peer has the validation logic on their machine — invalid data is treated as intentionally malicious, and validators can author **Warrants** — cryptographic proofs of bad behavior — against offending agents.

The [Warrants system (introduced in v0.6.0, November 2025)](https://blog.holochain.org/2025-at-a-glance-landing-reliability/) enables Byzantine agent isolation without global consensus: when an agent cheats (forks their chain, double-publishes), their own signed records expose the deceit. Warrants propagate to isolate the offender — a ["biological defense mechanism rather than a legal one"](https://www.reddit.com/r/holochain/comments/1pciogw/why_holochain_060_doesnt_need_global_consensus_to/).

**Current roadmap status** (from [Holochain official roadmap](https://www.holochain.org/roadmap/)):

| Version | Status | Key Features |
|---|---|---|
| 0.5.0 | Released April 2025 | **Kitsune2** networking rewrite, Wind Tunnel testing |
| 0.6.0 | Released November 2025 | Warrants, memproof security, coordinator updates |
| 0.6.1 | In progress (64%) | Performance improvements, per-app networking |
| 0.7.x | In progress (37%) | Data model consistency, HDK stability, DNA migration |

**Kitsune2** (networking layer rewrite) fixed the most significant reliability problems: [DHT synchronization that previously took 30+ minutes now works reliably](https://blog.holochain.org/2025-at-a-glance-landing-reliability/). **Wind Tunnel** (released January 2026) enables automated scale testing across arbitrary node counts. As of [March 2026](https://www.reddit.com/r/holochain/comments/1rz5k16/holochain_week_in_review_march_1420_2026/), the team delivers 30 story points per sprint with the Unyt pricing oracle launching as one of the first production-grade hApps.

**Membrane patterns** are Holochain's mechanism for flexible trust boundaries: a membrane proof can be an invite code, a signed credential, a proof of stake, or [any verifiable claim gating network access](https://holochain-gym.github.io/concepts/membranes-happs/). This enables graduated trust — agents move through membrane layers as they build track records — without central gatekeeping.

**Limitations:** Still beta; APIs change between major versions. No native token/incentive layer (intentional design choice). Entry size limit of 4MB. Small production hApp ecosystem. [DHT sharding remains behind experimental flags](https://blog.holochain.org/2025-at-a-glance-landing-reliability/).

### IPFS, IPLD, and Content-Addressable Storage

[IPFS](https://ipfsfoundation.org/content-addressing-2025-in-review/) provides tamper-evident, location-independent storage for agent artifacts. Content addressing means any agent requesting a CID gets exactly that content regardless of provider, enabling local verification without source trust. The [DASL (Decentralized Addressable Storage Layer) initiative](https://ipfsfoundation.org/content-addressing-2025-in-review/) improved cross-implementation interoperability, with an IETF Internet Draft submitted covering CIDs and DRISL.

For agent coordination, [Verified Fetch enables trustless browser retrieval](https://ipshipyard.com/blog/2024-shipyard-improving-ipfs-on-the-web/), Service Worker Gateways allow decentralized web app delivery, and WebRTC browser transport enables browser-to-browser content transfer. However, IPFS provides **no native persistence guarantees** — a [2025 IFIP study](https://tma.ifip.org/2025/wp-content/uploads/sites/14/2025/06/tma2025_paper16.pdf) found peer availability declined from 60% to 40%, with 50% of peers online for less than 4 days.

[Filecoin](https://filecoin.io/blog/posts/filecoin-in-2025-year-in-review/) addresses persistence through economic incentives: Proof-of-Replication and Proof-of-Spacetime cryptographically verify storage, with the network reaching 3.0 EiB capacity at ~$0.19/TB/month. The F3 (Fast Finality) upgrade in 2025 made Filecoin viable for applications needing faster confirmation. For agent provenance logs and model checkpoints, Filecoin provides the strongest decentralized persistence guarantees.

[IPLD (InterPlanetary Linked Data)](https://ipshipyard.com/blog/2024-shipyard-improving-ipfs-on-the-web/) enables content-addressed graphs where any node can reference any other by CID, including cross-protocol references to Ethereum, Bitcoin, or Git — useful for verifiable agent provenance chains.

### AD4M: Agent-Centric Composition Across Networks

[AD4M (Agent-centric Distributed Application Meta-ontology)](https://docs.ad4m.dev), developed by Coasys, provides a spanning layer atop existing protocols (Holochain, IPFS, HTTP, Solid) with a unified agent-centric abstraction. Each AD4M instance is sovereign, identified by a DID, running locally on the user's device with cryptographic keys, subjective knowledge graphs (Perspectives), and protocol adapters (Languages).

**The three core primitives:**

- **Agents**: Sovereign instances identified by DIDs (`did:key:z6Mk...`), signing every action with their private key, hosting a GraphQL API at localhost for local applications.
- **Languages**: Protocol abstractions that define how agents express, store, and share data — each identified by a content-addressed hash. Any existing protocol can be wrapped in a Language ([HTTP, IPFS, Holochain DHT, Solid, ActivityPub](https://coasys.org/adam)), enabling protocol-agnostic interoperability.
- **Links (RDF Triples)**: Building blocks of Perspectives — `<subject> <predicate> <object>` — each signed with cryptographic provenance and globally addressable across Languages.

**[Social DNA](https://docs.ad4m.dev)** encodes community rules in Prolog, evaluated locally by each agent without central arbiters. Different Neighbourhoods can have different rules (pluralistic alignment) while agents maintain a single identity across all contexts. This represents "soft law" — behavioral constraints enforced through protocol logic rather than external authority.

[AD4M v0.10.1 (February 2025)](https://github.com/coasys/ad4m/releases) integrates local AI inference — ADAM apps can run LLMs, transcription, and embedding locally (DeepSeek, Qwen, Whisper, Ollama-compatible models with CUDA/Metal GPU support), preserving privacy while enabling AI-augmented coordination.

**Limitations:** Pre-1.0; Windows binary not yet available. Synergy Engine (global semantic network) not yet production-ready. Small developer community. The ambitious spanning-layer vision depends on critical mass of Language implementations.

### Supporting Infrastructure

**[libp2p](https://libp2p.io)** provides the modular networking substrate used by IPFS, Ethereum, Polkadot, Filecoin, and Holochain's tx5 transport. It supports transport abstraction (TCP, WebSockets, QUIC, WebRTC, WebTransport), Kademlia DHT for peer discovery, and GossipSub for publish-subscribe messaging — the foundational plumbing for agent discovery and communication without central registries.

**[GunDB](https://en.wikipedia.org/wiki/GUN_(graph_database))** provides offline-first, real-time, CRDT-based graph databases running natively in browsers — useful for agent capability registries and coordination logs without server infrastructure.

**[Ceramic Network](https://ceramic.network)** enables mutable, DID-authenticated data streams anchored to Ethereum — suitable for portable agent reputation records and cross-application coordination state, though Ethereum anchoring creates gas costs and latency.

**[Nostr](https://arxiv.org/html/2402.05709v2)** provides key-sovereign, censorship-resistant messaging through relay networks. Agents can publish intents and capabilities as signed events and subscribe to relevant event kinds — maximally simple but with [relay centralization pressure and no persistence guarantees](https://dev.to/jurjendevries/2025-the-year-of-decentralization-how-nostr-will-make-you-a-standout-developer-5f5l).

**[Matrix](https://en.wikipedia.org/wiki/Matrix_(protocol))** provides federated real-time communication with DAG-based room state, E2E encryption, and enterprise-grade reliability (deployed for Germany's healthcare system and France's government communications). Agent rooms with structured event types and power-level permissions offer richer coordination than Nostr, at the cost of server infrastructure.

**[SSB (Secure Scuttlebutt)](https://ssbc.github.io/scuttlebutt-protocol-guide/)** offers fully offline-capable, identity-centric gossip with append-only signed feeds — agents publish decision logs and coordinate via follow-graph-mediated gossip. [PZP (Pico Zot Protocol)](https://autonomous.zone/@pzp) addresses SSB's scalability limitations.

**[KERI (Key Event Receipt Infrastructure)](https://trustoverip.github.io/kswg-keri-specification/)** provides the strongest decentralized public key infrastructure: self-certifying identifiers, append-only key event logs with pre-rotation, and witness-based accountability — all without dependency on DNS, certificate authorities, or blockchains. For AI agent identity, KERI enables non-repudiable action signing, secure key rotation, and [verifiable delegation chains](https://trustoverip.github.io/kswg-keri-specification/) from human principals to agent AIDs.

### Value Alignment Design Patterns

The fundamental insight across these technologies is that **value alignment can be embedded in protocol rules** rather than enforced by central authorities. As an [arxiv analysis on decentralized AI architecture](https://arxiv.org/html/2506.09656v1) argues, decentralized architecture inherently embodies values opposing power monopolies — the choice of architecture is itself a value statement.

**Cryptographic accountability** stacks identity binding (KERI/DIDs), action signing (all protocols), tamper-evident logs (SSB/Holochain source chains), content addressing (IPFS/IPLD), and [witnessed receipts (KERI/Holochain warrants)](https://arxiv.org/pdf/2512.17538) to create non-repudiable audit trails without central authorities.

**Membrane patterns** enable [graduated trust boundaries](https://www.socialroots.io/freedom-vs-order-the-decentralized-design-dilemma/) — composable, layered access control where agents move through trust levels as they build track records. As [SocialRoots (October 2025)](https://www.socialroots.io/freedom-vs-order-the-decentralized-design-dilemma/) argues: "Pure trustlessness is appropriate for censorship resistance but counterproductive for collaborative multi-agent systems. What if they need better 'membranes'?"

**Social DNA as soft law** — AD4M's Prolog-based behavioral rules enforce community norms locally without central arbiters. Combined with Holochain's intrinsic data validity (agents that produce invalid data are rejected by peers), this creates protocol-level alignment where rule-breaking makes agents unable to participate rather than merely punished.

## Automation Techniques to Reduce Human Bottlenecks

### CI/CD for Decentralized Agent Systems

Testing multi-agent systems is fundamentally harder than testing monolithic software because behaviors emerge from interactions. [Protocol-graph testing](https://pmc.ncbi.nlm.nih.gov/articles/PMC4385681/) transforms agent interaction protocols into directed graphs and derives test paths for coverage, while [property-based testing for MAS](https://dl.acm.org/doi/10.5555/3306127.3331931) declares behavioral properties and lets generators explore the interaction space.

[MAS-ProVe (February 2026)](https://arxiv.org/pdf/2602.03053) tested whether process-level verification — verifying agent reasoning trajectories, not just outputs — consistently improves LLM-based agent performance. The result was sobering: process-level verification "does not consistently improve performance and frequently exhibits high variance." LLM agent trajectories are stochastic and context-sensitive, making deterministic verification non-trivial.

**Reproducible builds** via [Nix](https://news.ycombinator.com/item?id=43963747) (>90% reproducibility across 80,000+ packages) eliminate environment inconsistency in distributed development. [Lila (January 2026)](https://arxiv.org/html/2601.20662v1) adds decentralized reproducibility monitoring, collecting 150,000+ attestations to identify previously unknown reproducibility regressions. Holochain's [Wind Tunnel](https://blog.holochain.org/2025-at-a-glance-landing-reliability/) provides purpose-built scale testing for distributed agent applications.

No mature, purpose-built CI/CD framework exists for decentralized agent behavior testing as of 2026 — this remains an open engineering gap.

### AI-Assisted Development: Current State and Limits

**Cursor** has become the fastest-growing developer tool in history — [$2B+ ARR, 1M+ daily active users, $29.3B valuation](https://www.getpanto.ai/blog/cursor-ai-statistics) — but is centralized and proprietary. **Devin** (Cognition AI) markets itself as autonomous but [completed only 3 of 20 complex tasks (15%) in independent testing](https://trickle.so/blog/devin-ai-review). **SWE-Agent** (open-source, MIT license) achieves [65% on SWE-bench Verified with just 100 lines of Python](https://www.swebench.com) — demonstrating that open-source approaches can match or exceed proprietary systems on narrow benchmarks.

The honest assessment: AI coding agents are transforming well-specified, contained tasks (migrations, API integrations, test generation) but fail at [complex debugging, visual reasoning, dependency conflict resolution, and multi-step root cause analysis](https://devin.ai/agents101). Real-world complex task completion rates remain ~15% without human assistance. For decentralized development, the limitation is deeper — all major coding agents rely on centralized LLM API endpoints, creating fundamental tension with decentralized principles.

### Multi-Agent Development Frameworks

| Framework | Architecture | Core Strength | Production Readiness | Key Limitation |
|---|---|---|---|---|
| **[LangGraph](https://gurusup.com/blog/best-multi-agent-frameworks-2026)** | Graph-based state machine | Checkpointing, debugging, explicit control | High (production standard) | Steep learning curve; single-machine |
| **[CrewAI](https://www.insightpartners.com/ideas/crewai-scaleup-ai-story/)** | Role-based agents | Fastest prototyping; 1.4B automations | Medium-High | 10+ min execution; limited observability |
| **[AutoGen v0.4](https://www.leanware.co/insights/auto-gen-vs-langgraph-comparison)** | Actor/message-passing | Distributed horizontal scaling | Medium | Documentation quality; verbose setup |
| **[MetaGPT](https://thirdeyedata.ai/technologies/metagpt)** | Team simulation (SOPs) | Full lifecycle automation | Low | Breaks on ambiguous requirements |

**None of these frameworks are meaningfully decentralized.** All use centralized orchestration, centralized LLM APIs, centralized state, and no cryptographic verification of agent outputs. The [Agent Communication Protocol (ACP)](https://semanticscholar.org/paper/ba14a755) proposes federated orchestration with decentralized identity and zero-trust security, while [AgentFlow](https://techrxiv.org/1292337) introduces a resilient cloud-edge framework with publish-subscribe messaging for many-to-many service elections. The [MAMA framework](https://arxiv.org/abs/2412.19256) combines a Verifiable Reputation Ledger with Expertise-Driven Routing. These remain research proposals, not production systems.

### Automated Governance: PR Bots, SourceCred, and Beyond

PR automation has matured significantly: [GitHub processes 43.2 million PRs per month](https://www.codeant.ai/blogs/top-pull-request-automation-tools), and AI-generated PRs have a [32.7% acceptance rate vs. 84.4% for human PRs](https://www.codeant.ai/blogs/top-pull-request-automation-tools). [Graphite Agent](https://graphite.com/guides/best-ai-pull-request-reviewers-2025) delivers instant AI code reviews in under 90 seconds with a 55% code-change rate when issues are flagged — higher than human reviewers at 49%.

**SourceCred** — the ambitious attempt to algorithmically measure and reward open-source contributions via PageRank-style algorithms — is [effectively discontinued](https://sourcecred.io/docs/). Gaming the algorithm proved persistent, and the project has no active releases. Alternatives with more staying power include [Coordinape](https://wiki.p2pfoundation.net/SourceCred) (peer-to-peer allocation circles) and reputation systems embedded in DAO toolkits like [Colony.io](https://docs.colony.io/develop/dev-learning/reputation/).

### Coordination Overhead Reduction

**Asynchronous coordination patterns** are critical for reducing bottlenecks. [Event-Triggered Control (ETC)](http://scis.scichina.com/en/2025/151201.pdf) reduces communication by triggering only on significant state changes. [AutoGen v0.4's actor model](https://www.leanware.co/insights/auto-gen-vs-langgraph-comparison) enables truly asynchronous agent interactions with horizontal scaling. [DXOS's ECHO with Automerge CRDTs](https://www.dxos.org) enables real-time multi-party collaboration without server coordination.

**Intent-centric architecture**, pioneered by [Anoma](https://anoma.net/blog/an-introduction-to-intents-and-intent-centric-architectures), separates user goals from execution mechanics. Users declare desired end states; specialized solvers compute execution paths. This maps naturally onto multi-agent systems — rather than orchestrators specifying agent assignments, a declarative intent layer allows agents to self-organize around fulfilling goals. The limitation is designing [incentive-compatible solver markets resistant to MEV-style extraction](https://www.mexc.co/en-GB/crypto-pulse/article/anoma-s-intent-centric-architecture-55885).

**Holacracy for agents** maps organizational governance directly: circles become agent groups, governance processes become DAO voting, and lead links become orchestrator agents with bounded scope. [Research confirms](https://kangwooklee.com/talks/2026_03_BLISS/bliss_seminar_monograph.html) that role separation in multi-agent systems prevents confirmation bias — the same LLM doing both coding and review would be biased, matching holacracy's principle of authority boundaries.

[Gossip protocols for agentic AI](https://arxiv.org/abs/2512.03285) enable context-rich state propagation and resilient coordination under uncertainty, with O(n log n) message complexity — substantially better than the O(n²) of direct communication approaches.

### The Silo-Bench Problem: Scaling Limits of Multi-Agent Coordination

[Silo-Bench (March 2026)](https://arxiv.org/html/2603.01045v1) provides the most rigorous empirical evidence for the coordination overhead wall. Key findings:

- Even at **team size k=2**, multi-agent systems lose **15–49%** of single-agent performance due to coordination overhead
- At **k=50**, coordination overhead **eliminates all parallelization gains** entirely for complex tasks
- Spontaneous leader emergence — often assumed beneficial — actively hurts performance at scale because the aggregator agent gets overwhelmed

This is a critical constraint for anyone designing large decentralized agent systems. The implication: asynchronous coordination should minimize total inter-agent communications, not just make them non-blocking. [Cognizant's India AI Summit](https://cognizant.com) found that practitioners instinctively gravitate toward hierarchical coordination over flat/decentralized swarms — an empirical validation that humans recognize the coordination overhead problem intuitively.

## Case Studies

### Autonolas (Olas) — Most Credible Implementation

**What they attempted:** A protocol for decentralized autonomous agent services combining off-chain AI logic with on-chain coordination, governance, and economic incentives using the [Open Autonomy framework](https://github.com/valory-xyz/open-autonomy).

**What actually worked:** Olas has the most concrete production metrics: [9.9 million+ agent-to-agent transactions](https://olas.network/timeline), 8.8M+ through the Mech Marketplace specifically. [Pearl v1](https://olas.network/blog/olas-the-2021-origin-story-of-one-of-the-first-crypto-ai-agent-projects-now-scaling-past-9-9-million-a2-a-transactions) (October 2025) launched as the first self-custodial AI Agent App Store with Web2 UX. Agents use Tendermint-like consensus for off-chain coordination, [Proof of Usefulness for code contributions](https://olas.network/timeline), and x402 integration for agent-to-agent payments. Multiple security audits completed; DAO governance operational.

**What failed:** OLAS token trading at ~$0.03 as of February 2026. Developer adoption concentrated in DeFi (prediction markets, DEX operations) rather than the broader vision. Whether the staking model is self-sustaining at current prices is unclear.

**Honest assessment:** Most honest executor in the space — published documentation, real audits, verifiable on-chain metrics. The 9.9M A2A transactions are genuine. The question is whether decentralized agent coordination escapes DeFi into broader automation.

### ElizaOS — Open-Source Agent Traction

**What they attempted:** Originally AI16Z, a crypto-native AI agent framework rebranded as [elizaOS](https://elizaos.ai) in late 2025, positioning as an "Agentic Operating System."

**What actually worked:** The open-source Eliza framework achieved genuine traction — [50,000+ autonomous agents](https://coinmarketcap.com/cmc-ai/elizaos/latest-updates/) across Solana, Ethereum, and Base. Plugin ecosystem (npm-based hot-swappable capabilities), Stanford partnership for trust frameworks, [Worlds/Rooms architecture](https://elizaos.ai), and integration with Chainlink CCIP for oracle access. Framework v1.4.4 represents significant architecture improvement.

**What failed:** The AI VC fund mechanism was ["more meme than genuine autonomous investment management"](https://www.panewslab.com/en/articles/atump910). Token depreciation despite rebrand. Decentralization claims are weak — orchestration, marketplace, and treasury remain controlled by Eliza Labs.

**Current status:** Alive and active (March 2026). Most widely adopted open-source AI agent framework in the Web3 space by contributor count. Commercial products (Eliza Cloud, agent marketplace) in active development.

### SingularityNET and the ASI Alliance

**What they attempted:** A [decentralized marketplace](https://singularitynet.io) for AI services with AGIX tokens for payments, staking, and governance, pursuing long-term AGI through the OpenCog Hyperon framework.

**What actually worked:** Functional AI marketplace with real services, [40 active partnerships](https://singularitynet.io/singularitynet-latest-ecosystem-updates-january-2025/), active Python SDK development, and HackIndia 2025 sponsorship (25,000 participants).

**What failed:** The ASI Alliance — a March 2024 token merger with Fetch.ai and Ocean Protocol — fractured when [Ocean Protocol exited in October 2025](https://blog.oceanprotocol.com/ocean-protocol-foundation-withdraws-from-the-artificial-superintelligence-alliance-4619c4604ea3), citing diverging visions. AGIX trades ~$0.06-0.07 (down 90%+ from ATH). Marketplace services are mostly research demos, not production-grade. Governance participation remains low.

**Honest assessment:** A genuine research organization making real but slow progress. The marketplace concept is technically sound but faces the fundamental challenge of competing with AWS Lambda and Hugging Face Spaces for developer attention.

### Fetch.ai — First Agent-to-Agent Payments

**What they attempted:** An autonomous economic agent ecosystem where AI agents discover, negotiate, and transact without human intermediation.

**What actually worked:** [The world's first AI-to-AI payment (December 2025)](https://www.linkedin.com/pulse/fetchai-announces-worlds-first-ai-to-ai-payment-real-world-transactions-qcroe) — two agents coordinated to book a dinner reservation and complete payment via Visa/USDC/FET while both users were offline. ASI:Chain DevNet launched November 2025.

**What failed:** [Financial difficulties led to administration and acquisition by Assembl.ai](https://liora.io/en/all-about-fetch-ai). FET declined 75% in 2024. The gap between "agents optimizing the European power grid" and "agent books dinner" is vast. Legal disputes with Ocean Protocol ongoing.

### Ocean Protocol — Compute-to-Data

**What they attempted:** A decentralized data exchange with privacy-preserving [Compute-to-Data (C2D)](https://blog.oceanprotocol.com/ocean-protocol-product-update-2025-half-year-check-in-49adcf092d87) — sending algorithms to data rather than data to algorithms.

**What actually worked:** C2D is technically innovative and addresses GDPR/HIPAA constraints. Ocean Nodes C2D.2 launched with VS Code integration. [GPU partnerships added ~2,000 high-performance GPUs](https://blog.oceanprotocol.com/ocean-protocol-q4-2025-update-9e275335d19b). 1.4 million nodes globally.

**What failed:** [ASI Alliance exit (October 2025)](https://blog.oceanprotocol.com/ocean-protocol-foundation-withdraws-from-the-artificial-superintelligence-alliance-4619c4604ea3) — a major governance failure for what was supposed to be a unified coalition. No self-sustaining data marketplace. C2D usage on meaningful private datasets remains rare.

### Colony.io — Reputation-Weighted Organization

**What they attempted:** On-chain organizational infrastructure with [reputation-based governance](https://blog.colony.io/what-is-reputation-based-governance), domain-specific reputation, temporal decay, and [lazy consensus](https://blog.colony.io/new-feature-simple-decisions/) for routine decisions.

**What actually worked:** The smart contract architecture works. The reputation system — non-transferable, earned through task completion, with domain-specific scoring — is a [genuinely novel mechanism](https://docs.colony.io/develop/dev-learning/reputation/). Flexible governance mixing consensus, voting, and hierarchical authority within sub-circles.

**What failed:** Never achieved mainstream DAO adoption. The full governance model is sophisticated enough that organizations default to simpler tools. The concept is valuable but finding the killer use case has proven elusive.

### Gitcoin — Quadratic Funding at Scale

**What they attempted:** Applying [quadratic funding](https://impact.gitcoin.co) — amplifying contributions based on breadth of support — to fund open-source public goods.

**What actually worked:** [**$50M+ distributed**, 3,715 projects funded, 270,000 unique supporters](https://impact.gitcoin.co), $28.2B combined peak grantee market cap (including Optimism, Uniswap, WalletConnect). [GG24 (October 2025)](https://gitcoin.co/case-studies/gg24-first-funding-round-of-gitcoin-3-0) distributed $1.8M using the new Domain Allocator model combining quadratic funding, conviction voting, MACI private voting, and hypercerts. Gitcoin Passport (identity layer) was sold for $10M.

**What failed:** [GTC token ~$0.10](https://gov.gitcoin.co/t/gitcoins-governance-strategy-for-2025/19845). Sybil attacks without reliable identity layers remain an ongoing challenge. Governance was self-described as "hectic, confusing and unreliable."

**Honest assessment:** The clearest success story in this landscape — solved a real coordination problem with demonstrably effective mechanisms at scale.

### Holochain Ecosystem Projects

**What they attempted:** [Truly peer-to-peer applications](https://blog.holochain.org/2025-at-a-glance-landing-reliability/) without blockchain consensus. 2025 focused on reliability over features — "making what we have actually work reliably."

**What actually worked:** Kitsune2 networking rewrite fixed peer discovery. Validation pipeline fixed. Warrants system operational. [Wind Tunnel reached production readiness](https://blog.holochain.org/2025-at-a-glance-landing-reliability/). Active applications include Moss, Carbon Farm Network, Arkology Studio, HummHive, and [Unyt (pricing oracles, launched March 2026)](https://www.reddit.com/r/holochain/comments/1rz5k16/holochain_week_in_review_march_1420_2026/).

**What failed:** After 7+ years, no mainstream adoption. HoloFuel/HOT token swap delayed repeatedly. DHT sharding at scale remains experimental. [Holo hosting economics proved difficult](https://holo.host/blog/2025-year-in-review-the-year-we-built-the-edge-XqpCNKmMRVh/) — pivoting to Edge Node.

**Honest assessment:** Most technically sophisticated approach to P2P application infrastructure. The 2025 reliability focus was the right call. The gap between technical capability and adoption remains the central challenge.

### Open-Source Frameworks: LangGraph, CrewAI, AutoGen

[LangGraph](https://gurusup.com/blog/best-multi-agent-frameworks-2026) is the production standard (27,100 monthly searches, checkpointing, time-travel debugging). [CrewAI Enterprise](https://www.insightpartners.com/ideas/crewai-scaleup-ai-story/) reports 1.4B automations but [practitioners note 10+ minute execution times and limited production credibility](https://www.reddit.com/r/AI_Agents/comments/1l6rw2n/whos_using_crewai_really/). [AutoGen v0.4](https://www.leanware.co/insights/auto-gen-vs-langgraph-comparison) rebuilt on actor model enables horizontal scaling. [MetaGPT](https://thirdeyedata.ai/technologies/metagpt) is best for bounded projects. None are meaningfully decentralized.

### Case Study Synthesis Table

| Project | What Worked | What Failed | Status | Key Lesson |
|---|---|---|---|---|
| **[Autonolas](https://olas.network)** | 9.9M A2A transactions; Mech Marketplace | DeFi-concentrated; token collapse | Alive, shipping | On-chain metrics matter; vision vs. reality |
| **[ElizaOS](https://elizaos.ai)** | 50k+ agents; strong OSS adoption | AI VC was theater; weak decentralization | Alive, active | Framework value > token speculation |
| **[SingularityNET](https://singularitynet.io)** | Functional marketplace; 40 partnerships | ASI Alliance fracture; low adoption | Alive, diminished | Token mergers fragment on identity |
| **[Fetch.ai](https://liora.io/en/all-about-fetch-ai)** | First AI-to-AI payment demo | Financial difficulties; vision >> reality | Alive, struggling | Demos ≠production deployments |
| **[Ocean Protocol](https://blog.oceanprotocol.com)** | C2D innovation; GPU partnerships | ASI Alliance exit; no marketplace flywheel | Alive, pivoting | Privacy tech needs demand-side |
| **[Colony.io](https://colony.io)** | Reputation governance mechanism | No killer use case; limited adoption | Alive, niche | Solutions need problems |
| **[Gitcoin](https://impact.gitcoin.co)** | $50M+ distributed; QF proven at scale | GTC decline; governance instability | Alive, impactful | QF works; token value ≠impact |
| **[Holochain](https://www.holochain.org/roadmap/)** | 2025 reliability; Warrants system | 7+ years, not mainstream; HoloFuel delays | Alive, maturing | Technical excellence ≠adoption |
| **[LangGraph](https://gurusup.com/blog/best-multi-agent-frameworks-2026)** | Production standard; checkpointing | Single-machine; not decentralized | Dominant | Centralized works until it doesn't |

## Unsolved Problems and Legitimate Limitations

### The Trust Gap in Multi-Agent Coordination

The [Consensys/MetaMask letter to NIST](https://consensys.io/blog/ai-agents-are-handling-real-money) frames agents as "software with delegated authority" requiring scoped, revocable delegations. The proposed [ERC-8004 trust infrastructure](https://consensys.io/blog/ai-agents-are-handling-real-money) and x402 protocol for agent payments represent early standards, but the gap between delegating authority and verifying that authority was exercised correctly remains wide. [ISACA's 2026 findings](https://isaca.org) are alarming: only **14.4% of enterprises** get full security approval before deploying AI agents, with an 82:1 ratio of autonomous agents to human employees and 45.6% using shared API keys.

### Scalability Wall: Coordination Overhead at Scale

[Silo-Bench](https://arxiv.org/html/2603.01045v1) demonstrates that coordination overhead is not merely a theoretical concern but an empirically measured wall. At k=50 agents, all parallelization gains evaporate. This finding constrains the entire field: architectures that require inter-agent communication scaling as O(n²) are fundamentally limited. Only approaches with sublinear communication — [stigmergic coordination](https://www.distributedthoughts.org/digital-pheromones-what-ants-know-about-agent-coordination/), [CRDTs (zero coordination at runtime)](https://crdt.tech), and [gossip protocols (O(n log n))](https://arxiv.org/abs/2512.03285) — can plausibly scale to hundreds or thousands of agents.

### LLM Non-Determinism and Formal Verification

[MAS-ProVe (2026)](https://arxiv.org/pdf/2602.03053) found that process-level verification does not consistently improve LLM agent performance. The fundamental difficulty: LLM outputs are probabilistic, context-sensitive, and non-deterministic. Formal verification techniques developed for rule-based agents with well-defined action spaces cannot be directly applied. [Imperial College's verification frameworks for open MAS](https://pkouvaros.github.io/publications/AAMAS19-K+/paper.pdf) face state-space explosion even for classical agents — for LLM agents, the state space is effectively unbounded.

### Identity, Accountability, and Delegation in Open Systems

The accountability stack — [KERI-based identity, action signing, tamper-evident logs, content addressing, and witnessed receipts](https://arxiv.org/pdf/2512.17538) — exists conceptually but lacks integrated production implementations. The critical missing piece is **verifiable delegation chains**: when an AI agent acts on behalf of a human principal, the authorization chain must be cryptographically recorded and auditable. [KERI's key event delegation](https://trustoverip.github.io/kswg-keri-specification/) and W3C DID delegation support this in theory, but production tooling is immature.

### Economic Sustainability of Decentralized Agent Networks

Every decentralized agent token (OLAS, GTC, FET, AGIX, OCEAN) has declined dramatically in value. Decentralized token marketplaces for services — [SingularityNET, Ocean, Fetch.ai](https://singularitynet.io) — have not achieved self-sustaining network effects after 5–7 years of operation. The fundamental tension: making decentralized agent coordination economically viable requires either token-based incentives (which are volatile and speculative) or traditional payment infrastructure (which re-centralizes). [Autonolas's Proof of Usefulness](https://olas.network/timeline) for code contributions represents one approach, but sustainability at current token prices remains unproven.

### The Centralization Gravity Problem

Despite the decentralized vision, centralization pressure is persistent and multidimensional. All major agent frameworks use centralized LLM API endpoints (OpenAI, Anthropic). Nostr users [concentrate on popular relays](https://arxiv.org/html/2402.05709v2). IPFS peer availability [declines without economic incentives](https://tma.ifip.org/2025/wp-content/uploads/sites/14/2025/06/tma2025_paper16.pdf). Even open-source projects like ElizaOS maintain [centralized control over orchestration and treasury](https://coinmarketcap.com/cmc-ai/elizaos/latest-updates/). [AD4M's local AI integration](https://github.com/coasys/ad4m/releases) (running models locally via Kalosm) represents a partial answer, but current local models significantly underperform centralized API models.

### Security: OWASP Top 10 for Agentic Applications

The [OWASP Top 10 for Agentic Applications (February 2026)](https://owasp.org) catalogs systemic risks including prompt injection, insufficient authorization, excessive agency, and supply chain vulnerabilities. When combined with [ISACA's finding that 45.6% of enterprises use shared API keys](https://isaca.org) for agent access, the security posture of deployed agent systems is deeply concerning. Decentralized systems add attack surfaces: compromised validators in BFT systems, CRDT metadata manipulation, and Sybil attacks on reputation systems.

### The Governance Implementation Gap

The governance mechanism literature is rich — conviction voting, quadratic voting, futarchy, reputation-based governance — but production implementations lag dramatically. [Gitcoin's own governance](https://gov.gitcoin.co/t/gitcoins-governance-strategy-for-2025/19845) self-described as "hectic, confusing and unreliable." Token-weighted DAOs suffer [plutocratic capture](https://www.chainscorelabs.com/en/blog/network-states-and-pop-up-cities/network-state-governance-models/why-reputation-based-governance-will-outperform-token-weighted-voting). Multi-token alliance mergers (ASI Alliance) [fracture on community identity](https://blog.oceanprotocol.com/ocean-protocol-foundation-withdraws-from-the-artificial-superintelligence-alliance-4619c4604ea3). The gap between governance theory and governance practice remains wide enough that most production decentralized systems operate with de facto centralized decision-making.

## Conclusions and Actionable Recommendations

### What the Evidence Actually Supports

The evidence supports several specific claims while refuting others:

**Supported:**
- CRDTs provide [zero-coordination-overhead state synchronization](https://arxiv.org/pdf/2510.18893) with mathematical convergence guarantees, validated at [1,000+ nodes](https://dl.acm.org/doi/10.1145/3756907.3756910)
- [ACO-based agent routing scales to 1,000 concurrent agents](https://arxiv.org/html/2603.12933) with measurable performance gains (4.7x speedup)
- [Holochain's agent-centric model eliminates the global consensus bottleneck](https://pmc.ncbi.nlm.nih.gov/articles/PMC12251913/) with 4x lower publish latency than blockchain
- [Quadratic funding works at scale](https://impact.gitcoin.co) ($50M+ distributed, measurable ecosystem impact)
- [Reputation-based governance provides stronger Sybil resistance and expertise alignment](https://blog.colony.io/what-is-reputation-based-governance) than token-weighted voting
- [On-chain agent-to-agent coordination is technically viable](https://olas.network/timeline) (9.9M transactions)

**Not supported:**
- Naively adding more agents improves performance — [it does not beyond small teams](https://arxiv.org/html/2603.01045v1)
- Decentralized token marketplaces achieve self-sustaining network effects — [none have after 5–7 years](https://singularitynet.io)
- Process-level verification consistently improves LLM agent quality — [it does not](https://arxiv.org/pdf/2602.03053)
- Token mergers create unified ecosystems — [they fracture on identity](https://blog.oceanprotocol.com/ocean-protocol-foundation-withdraws-from-the-artificial-superintelligence-alliance-4619c4604ea3)

### A Layered Architecture for Decentralized Agent Orchestration

Based on the evidence, the recommended architecture is layered by concern:

1. **Trust Layer — Holochain:** DNA-based validation rules encode what valid agent behavior looks like. [Membrane proofs](https://holochain-gym.github.io/concepts/membranes-happs/) gate network access without central gatekeepers. [Warrants](https://blog.holochain.org/2025-at-a-glance-landing-reliability/) isolate malicious agents cryptographically. Source chains provide tamper-evident audit trails.

2. **State Synchronization — CRDTs:** Shared agent state (task queues, configuration, voting tallies) uses [CRDTs for zero-coordination-overhead convergence](https://arxiv.org/pdf/2510.18893). [CodeCRDT's TODO-claim protocol](https://arxiv.org/pdf/2510.18893) provides a validated pattern for concurrent agent work allocation.

3. **Discovery and Propagation — Gossip Protocols:** [Context-rich gossip](https://arxiv.org/abs/2512.03285) enables resilient agent discovery and state propagation with O(n log n) message complexity. [libp2p's GossipSub](https://libp2p.io) provides the production-validated networking substrate.

4. **Task Allocation — Contract Net Protocol:** [CNP's self-organizing auction dynamics](https://notes.muthu.co/2025/10/task-allocation-for-machine-teamwork-with-the-contract-net-protocol/) allocate work to the most suitable agents without centralized capability databases. For large-scale routing, [ACO with task-specific pheromone matrices](https://arxiv.org/html/2603.12933) scales to 1,000+ agents.

5. **Governance — Reputation-Weighted Mechanisms:** [Colony.io's model](https://docs.colony.io/develop/dev-learning/reputation/) — domain-specific, non-transferable, temporally decaying reputation — aligns governance weight with demonstrated competence. [Conviction voting](https://gitcoin.co/mechanisms/conviction-voting) for treasury allocation rewards sustained commitment. For high-stakes decisions, [futarchy](https://www.frontiersin.org/journals/blockchain/articles/10.3389/fbloc.2025.1650188/full) aggregates distributed information through prediction markets.

6. **Semantic Interoperability — AD4M:** [AD4M's Perspectives, Languages, and Links](https://docs.ad4m.dev) provide the spanning layer for cross-protocol agent communication. [Social DNA](https://docs.ad4m.dev) encodes community behavioral norms as executable Prolog rules, enabling pluralistic value alignment without central enforcement.

7. **Identity — KERI:** [Self-certifying identifiers with pre-rotation](https://trustoverip.github.io/kswg-keri-specification/) provide the strongest cryptographic root of trust for agent identity, independent of any specific infrastructure.

8. **Persistence — IPFS + Filecoin:** [Content-addressable storage](https://ipfsfoundation.org/content-addressing-2025-in-review/) for tamper-evident artifacts; [Filecoin's economic incentives](https://filecoin.io/blog/posts/filecoin-in-2025-year-in-review/) for cryptographically proven persistence.

This architecture separates concerns to avoid the one-size-fits-all trap. The key design principle: use the minimal coordination mechanism sufficient for each layer. Strong consensus (BFT) is reserved for financial settlement between untrusted agents. Everything else uses eventually consistent, coordination-free approaches.

### Open Questions for Further Research

1. **Coordination overhead curves**: Silo-Bench studied flat and spontaneously hierarchical teams. How do pre-designed hierarchical architectures (like [AgentNet++'s 23% improvement](https://arxiv.org/abs/2512.00614)) perform at k=50–100 scale? What is the empirical optimal team size for different task categories?

2. **Cross-membrane agent migration**: How should agent reputation and credentials transfer when an agent moves between Holochain networks (different DNAs) or AD4M Neighbourhoods? Verifiable credential portability is theoretically supported by KERI/DIDs but lacks production validation.

3. **Economic sustainability without tokens**: Can decentralized agent networks sustain themselves through service fees, subscription models, or [quadratic funding](https://impact.gitcoin.co) for public goods, without relying on volatile token economics?

4. **LLM non-determinism in consensus**: When LLM agents participate in BFT consensus, their non-deterministic outputs create novel failure modes distinct from classical Byzantine faults. What verification approaches can provide probabilistic guarantees?

5. **FLOSSIOULLK integration path**: Bridging Holochain's agent-centric trust (DNA validation, Warrants, membranes) with AD4M's semantic spanning layer (Perspectives, Languages, Social DNA) creates a viable foundation for the FLOSSIOULLK/ARF architecture. The critical engineering gap is bridging Holochain's Rust-based validation with AD4M's Deno/V8 Language runtime at scale, and validating the full stack under adversarial conditions using Wind Tunnel.

6. **Value alignment at the protocol layer**: The [argument that decentralized architecture inherently embodies anti-monopolistic values](https://arxiv.org/html/2506.09656v1) deserves rigorous testing. Does the choice of coordination mechanism measurably affect the values expressed by the resulting agent network? Can Social DNA encode ethical constraints that meaningfully constrain LLM agent behavior?

The field of decentralized agent orchestration is at an inflection point: the infrastructure components exist in beta form, coordination algorithms have demonstrated gains in controlled settings, and production deployments have validated specific patterns. The path from here to autonomous agent economies is not primarily a technology problem — it is a coordination problem, which is fitting given the subject matter.
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Fix character encoding corruption across the document.

The file contains widespread mojibake (—, –, â‰, O(n²)), which makes technical claims harder to read and can distort meaning in formulas/comparisons.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/research/Perplexity-Source-Agent-Orchestration-March2026.md` around
lines 5 - 454, The document contains widespread mojibake (e.g. sequences like
"—", "–", "â‰", "O(n²)") due to wrong/declared encoding; convert the file
to UTF-8 (ensure correct byte decoding), normalize Unicode (NFC), and run a
targeted find-and-replace for common corruptions replacing "—"→"—" (em dash),
"–"→"–" (en dash), "â‰"→"≠" or "≈" as context requires, and "O(n²)"→"O(n²)"
(and similar superscript/degree issues); verify math/inline formulas (e.g. the
pheromone update rule and tables) preserve intended symbols and run a
diff/spot-check on headings like the document title
"Perplexity-Source-Agent-Orchestration-March2026.md" and occurrences of "CRDTs",
"Holochain", "Silo-Bench" to confirm no semantic corruption.

Comment on lines +55 to +68
```
Agent A calls assert_triple({
subject: "holochain",
predicate: "is_a",
object: "distributed_framework",
confidence: 0.95
})

ASSERT: returns ActionHash (not error)
ASSERT: get(action_hash) returns Record with:
- entry content matching input
- action.author == Agent A's pubkey
- action.timestamp is recent (within 30s)
```
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🧹 Nitpick | 🔵 Trivial

Add language specifier to test pseudocode blocks.

The test pseudocode blocks (lines 55-68, 72-78, 82-87, 91-111, 115-118) lack language specifiers, which reduces readability and IDE support. While these are pseudocode rather than executable code, specifying a language improves clarity.

Recommendation: Add language specifiers to fenced code blocks, e.g.:

```typescript
Agent A calls assert_triple({
  subject: "holochain",
  ...
})
```
🧰 Tools
🪛 markdownlint-cli2 (0.21.0)

[warning] 55-55: Fenced code blocks should have a language specified

(MD040, fenced-code-language)

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/specs/phase0-substrate-bridge.spec.md` around lines 55 - 68, The fenced
pseudocode blocks in the spec (containing calls like assert_triple({...}),
ASSERT lines, references to ActionHash and get(action_hash)) should include a
language specifier to improve readability and editor support; update each
triple/assertion pseudocode block to start with a fenced code block marker that
names a language (e.g., ```typescript or ```pseudo) so IDEs/highlighters treat
the content appropriately, leaving the pseudocode unchanged otherwise.

Comment on lines +140 to +142
2. **Provenance retrieval helper** — The existing `query_triples` returns content but not full provenance (author, timestamp). Either:
- (a) Add a `get_triple_with_provenance` extern that returns `Record` metadata, or
- (b) Add author/timestamp fields to `TripleResult` struct
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

🧩 Analysis chain

🏁 Script executed:

#!/bin/bash
# Check current Rose Forest DNA for KnowledgeTriple and TripleResult definitions

echo "=== Checking KnowledgeTriple entry type ==="
rg -n "struct KnowledgeTriple" ARF/dnas/ -A 10

echo -e "\n=== Checking TripleResult struct ==="
rg -n "struct TripleResult" ARF/dnas/ -A 10

echo -e "\n=== Checking query_triples extern ==="
rg -n "fn query_triples" ARF/dnas/ -A 15

echo -e "\n=== Checking assert_triple extern ==="
rg -n "fn assert_triple" ARF/dnas/ -A 10

Repository: kalisam/FLOSS

Length of output: 5482


🏁 Script executed:

cat -n docs/specs/phase0-substrate-bridge.spec.md | head -60

Repository: kalisam/FLOSS

Length of output: 2859


🏁 Script executed:

sed -n '135,150p' docs/specs/phase0-substrate-bridge.spec.md

Repository: kalisam/FLOSS

Length of output: 778


Resolve provenance design choice in specification.

Lines 140-142 identify an unresolved design decision: query_triples returns content but not full provenance (author, timestamp), proposing two alternatives:

  • (a) Add get_triple_with_provenance extern returning Record metadata
  • (b) Add author/timestamp fields to TripleResult struct

Issue: Success criteria 2 and 3 require provenance verification, though they retrieve it via get by hash rather than through query_triples. The spec is implementable but leaves implementers to choose between two approaches with different ergonomic and architectural trade-offs.

Recommendation: Choose option (b) and document it. The codebase already tracks provenance in KnowledgeTriple (source, created_at fields), and extending TripleResult to include these aligns with the learnings that all entry types should include provenance fields for the provenance-first architecture.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/specs/phase0-substrate-bridge.spec.md` around lines 140 - 142, Choose
option (b): extend the TripleResult struct to include provenance fields
(author/source and timestamp/created_at) instead of introducing a separate
get_triple_with_provenance extern; update the spec text around query_triples to
state that Query results include provenance, add author/timestamp fields to
TripleResult, and ensure the mapping from existing KnowledgeTriple.source and
KnowledgeTriple.created_at to the new TripleResult fields is documented so
implementers can return provenance directly from query_triples.

Comment on lines +162 to +168
## 6. Definition of Done

- [ ] `substrate_bridge.test.ts` written with all 6 criteria as test cases
- [ ] All 6 tests pass against running Holochain conductor
- [ ] Results logged in `docs/governance/HARVEST_LOG.md` as substrate bridge validation
- [ ] ADR-2 (Holochain Substrate) updated from Proposed to Accepted (evidence: this test passing)

Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Definition of done includes critical dependency on unresolved gap.

The DoD checklist (lines 164-165) requires "All 6 tests pass against running Holochain conductor," but test criteria 2, 3, and 5 depend on provenance retrieval, which lines 140-142 acknowledge as not currently possible with existing externs.

Issue: This creates a circular dependency - the DoD cannot be achieved until the provenance retrieval gap is resolved, but the spec doesn't mandate that resolution.

Recommendation: Add a DoD item before line 164:

- [ ] Provenance retrieval method implemented (either get_triple_with_provenance extern or TripleResult extension)
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@docs/specs/phase0-substrate-bridge.spec.md` around lines 162 - 168, Add a new
Definition of Done checklist item that requires implementation of provenance
retrieval before running the substrate bridge tests: require a provenance
retrieval method (either the get_triple_with_provenance extern or an extension
to TripleResult) to be implemented and validated so that tests referenced by
substrate_bridge.test.ts (notably test criteria 2, 3, and 5) can pass; ensure
this DoD item is placed before the existing "All 6 tests pass against running
Holochain conductor" entry and mention that its completion is prerequisite
evidence for ADR-2 acceptance and HARVEST_LOG.md recording.

Copy link
Copy Markdown

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Warning

CodeRabbit couldn't request changes on this pull request because it doesn't have sufficient GitHub permissions.

Please grant CodeRabbit Pull requests: Read and write permission and re-run the review.

👉 Steps to fix this

Actionable comments posted: 3

🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@ARF/tests/tryorama/substrate_bridge.test.ts`:
- Around line 146-148: The test currently only asserts that record.entry exists
but never verifies that Bob received the actual triple; update the assertions to
compare record.entry.subject, record.entry.predicate, record.entry.object, and
record.entry.confidence against the original values Alice wrote (e.g., the
variables used when creating Alice's entry or an aliceEntry object) — for
example assert.strictEqual(record.entry.subject, expectedSubject),
assert.strictEqual(record.entry.predicate, expectedPredicate),
assert.strictEqual(record.entry.object, expectedObject), and
assert.strictEqual(record.entry.confidence, expectedConfidence) (or a numeric
tolerance check for confidence) so the test truly verifies replicated content
from Alice.
- Around line 77-81: Add a negative test that calls get_triple_record with a
non-existent or wrong-type hash and assert the expected failure (e.g., returns
null or throws a specific error) to ensure adversarial coverage; locate the
existing positive call using aliceCall("get_triple_record", hash) in
ARF/tests/tryorama/substrate_bridge.test.ts and add another case that supplies a
fakeHash (or a hash from another entry type) and asserts the negative outcome,
and replicate analogous negative-case additions for the other occurrences around
the blocks referenced (lines ~139-142 and ~334-341) so each positive path has a
corresponding adversarial test.
- Around line 27-30: The bundle path is being resolved from the test file path
instead of its directory; update how hAppPath is built by calling
path.dirname(fileURLToPath(import.meta.url)) and then passing that directory
into path.resolve with "../../../workdir/rose_forest.happ" so hAppPath is
resolved from the directory rather than from the file path (change the hAppPath
construction that currently uses fileURLToPath(import.meta.url) directly to use
path.dirname(fileURLToPath(import.meta.url)) first).
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: Organization UI

Review profile: ASSERTIVE

Plan: Pro

Run ID: 3dfc7880-0388-4bba-8837-24ca09283e87

📥 Commits

Reviewing files that changed from the base of the PR and between 36ea046 and 00d1478.

📒 Files selected for processing (2)
  • ARF/dnas/rose_forest/zomes/coordinator/src/lib.rs
  • ARF/tests/tryorama/substrate_bridge.test.ts

Comment on lines +27 to +30
const hAppPath = path.resolve(
fileURLToPath(import.meta.url),
"../../../workdir/rose_forest.happ"
);
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🔴 Critical

🧩 Analysis chain

🏁 Script executed:

#!/bin/bash
node - <<'NODE'
const path = require("node:path");

const testFile = path.resolve("ARF/tests/tryorama/substrate_bridge.test.ts");
console.log("current :", path.resolve(testFile, "../../../workdir/rose_forest.happ"));
console.log("expected:", path.resolve(path.dirname(testFile), "../../../workdir/rose_forest.happ"));
NODE

find . -name 'rose_forest.happ' -print

Repository: kalisam/FLOSS

Length of output: 1792


🌐 Web query:

Node.js path.resolve() file vs directory semantics

💡 Result:

Node.js path.resolve has identical semantics for files and directories. It resolves a sequence of paths or path segments into an absolute path string using the current working directory if needed. It performs pure string manipulation and normalization (handling ., .., and separators), without accessing the filesystem to check existence, type, or whether the path refers to a file or directory. Examples from official docs: path.resolve('/foo/bar', './baz') returns '/foo/bar/baz' path.resolve('/foo/bar', '/tmp/file/') returns '/tmp/file/' path.resolve('wwwroot', 'static_files/png/', '../gif/image.gif') returns full absolute path based on cwd The resulting path string is the same regardless of whether the final path points to a file, directory, symlink, or non-existent location. To distinguish file vs directory, use fs.stat or fs.lstat separately after resolving the path.

Citations:


Resolve the bundle path from the directory, not from the file.

fileURLToPath(import.meta.url) returns the file path including substrate_bridge.test.ts. When passed directly to path.resolve(), the relative segments ../../../ are resolved from the file path itself, causing the bundle to be looked up at ARF/workdir/rose_forest.happ instead of workdir/rose_forest.happ. Extract the directory first with path.dirname() before resolving the relative path.

Proposed fix
 const hAppPath = path.resolve(
-  fileURLToPath(import.meta.url),
+  path.dirname(fileURLToPath(import.meta.url)),
   "../../../workdir/rose_forest.happ"
 );
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
const hAppPath = path.resolve(
fileURLToPath(import.meta.url),
"../../../workdir/rose_forest.happ"
);
const hAppPath = path.resolve(
path.dirname(fileURLToPath(import.meta.url)),
"../../../workdir/rose_forest.happ"
);
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@ARF/tests/tryorama/substrate_bridge.test.ts` around lines 27 - 30, The bundle
path is being resolved from the test file path instead of its directory; update
how hAppPath is built by calling path.dirname(fileURLToPath(import.meta.url))
and then passing that directory into path.resolve with
"../../../workdir/rose_forest.happ" so hAppPath is resolved from the directory
rather than from the file path (change the hAppPath construction that currently
uses fileURLToPath(import.meta.url) directly to use
path.dirname(fileURLToPath(import.meta.url)) first).

Comment on lines +77 to +81
// Retrieve the full Record with provenance
const record = await aliceCall<HcRecord | null>(
"get_triple_record",
hash
);
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Add a negative get_triple_record scenario.

Every call here passes a known-good triple hash. The suite still needs at least one adversarial case for a missing hash or a hash from another entry type; otherwise a contract regression in get_triple_record can ship while this spec stays green. As per coding guidelines ARF/tests/**/*.{py,ts}: Test at multiple layers: Unit tests for individual functions, integration tests for multi-component coordination, system tests for end-to-end scenarios, reality validation with empirical measurement and adversarial testing.

Also applies to: 139-142, 334-341

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@ARF/tests/tryorama/substrate_bridge.test.ts` around lines 77 - 81, Add a
negative test that calls get_triple_record with a non-existent or wrong-type
hash and assert the expected failure (e.g., returns null or throws a specific
error) to ensure adversarial coverage; locate the existing positive call using
aliceCall("get_triple_record", hash) in
ARF/tests/tryorama/substrate_bridge.test.ts and add another case that supplies a
fakeHash (or a hash from another entry type) and asserts the negative outcome,
and replicate analogous negative-case additions for the other occurrences around
the blocks referenced (lines ~139-142 and ~334-341) so each positive path has a
corresponding adversarial test.

Comment on lines +146 to +148
// Verify content matches what Alice wrote
const entry = record!.entry;
assert.ok(entry, "Record should have entry data");
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

Criterion 3 never verifies the replicated triple content.

This only checks that record.entry exists. It never proves Bob fetched Alice’s subject, predicate, object, and confidence, so the test still passes if the wrong app entry comes back with valid provenance.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@ARF/tests/tryorama/substrate_bridge.test.ts` around lines 146 - 148, The test
currently only asserts that record.entry exists but never verifies that Bob
received the actual triple; update the assertions to compare
record.entry.subject, record.entry.predicate, record.entry.object, and
record.entry.confidence against the original values Alice wrote (e.g., the
variables used when creating Alice's entry or an aliceEntry object) — for
example assert.strictEqual(record.entry.subject, expectedSubject),
assert.strictEqual(record.entry.predicate, expectedPredicate),
assert.strictEqual(record.entry.object, expectedObject), and
assert.strictEqual(record.entry.confidence, expectedConfidence) (or a numeric
tolerance check for confidence) so the test truly verifies replicated content
from Alice.

Code changes (ternary confidence):
- Rust integrity: confidence validation [0,1] -> [-1,1] for KnowledgeTriple
  and KnowledgeEdge. Negative = away from truth, 0 = neutral, + = toward truth.
- JSON schemas: knowledge-triple.schema.json and knowledge-edge.schema.json
  updated to minimum: -1.0.
- Tests: 2 new cases (negative confidence valid, below -1.0 rejected).
- TERNARY_COMPATIBILITY.md: maps what can/cannot use signed gradients.
  Flagged: impact (magnitude), budget (quantity), embeddings (external models),
  Holochain Timestamp (external SDK type).
- Already ternary: connotation [-1,0,+1], cosine_similarity [-1,1].

FLOSSI U Founding Kit v1.6 (containing scope):
- FLOSSI U is the vision/curriculum layer; FLOSS repo is the nested
  implementation substrate. Separate ADR namespaces (U: 001-019, repo: 0-N).
- 19 curriculum ADRs, seed packet, art effects, reference docs.
- ADR-019 cross-references docs/specs/TERNARY_COMPATIBILITY.md for
  external interface constraints that prevent ternary adoption.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
return (subject, predicate, obj)

# Fallback patterns
is_a_pattern = r'(\S+(?:-\S+)*)\s+is\s+an?\s+([\w\s-]+?)(?:\s*$|[.,;!?])'

Check failure

Code scanning / CodeQL

Inefficient regular expression High

This part of the regular expression may cause exponential backtracking on strings starting with '!-' and containing many repetitions of '!-'.
obj = match.group(2).strip().replace(' ', '-')
return (subject, IS_A, obj)

improves_pattern = r'(\S+(?:-\S+)*)\s+improves(?:\s+upon)?\s+(\S+(?:-\S+)*)'

Check failure

Code scanning / CodeQL

Inefficient regular expression High

This part of the regular expression may cause exponential backtracking on strings starting with '!-' and containing many repetitions of '!-'.
if match:
return (match.group(1).strip(), IMPROVES_UPON, match.group(2).strip())

capable_pattern = r'(\S+(?:-\S+)*)\s+(?:can|is capable of)\s+(\w+)'

Check failure

Code scanning / CodeQL

Inefficient regular expression High

This part of the regular expression may cause exponential backtracking on strings starting with '!-' and containing many repetitions of '!-'.
kalisam and others added 3 commits April 4, 2026 13:52
Comparative analysis of Meta Harness, omo (oh-my-openagent), OMX/OMC,
and FLOSSI0ULLK MetaCoordinator. Proposes 5 integration seams (consensus-
gate hook, git worktree, LiteLLM optimization, KAIROS memory, OMX gateway
events), 18 patterns to ADOPT, 12 patterns to CONTRIBUTE, with
implementation priority table (Seam 1 first).

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
ADR-6 formalizes the four-system integration strategy across Meta Harness,
omo, OMX, and FLOSSI0ULLK. Seam 1 (consensus-gate hook) is implemented
first because it establishes the MCP channel that Seams 2-5 depend on.

Adds:
- docs/adr/ADR-6-four-system-integration.md (Proposed)
- docs/specs/consensus-gate.spec.md — Claim/Vote/Decision contract
- docs/specs/consensus-gate.schema.json — JSON Schema
- packages/orchestrator/claim_schema.py — dataclasses + invariants
- packages/orchestrator/consensus_gate.py — tally/decide/override + ADR writer
- packages/orchestrator/test_consensus_gate.py — 16 tests, all 6 spec vectors pass

Ternary consensus: {-1, 0, +1} votes => {APPROVED, DEFERRED, REJECTED,
OVERRIDDEN}. Single -1 vetoes. Substrate blast radius requires unanimous
+1 with no override path. Human override only on DEFERRED decisions for
non-substrate blast radii.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
- resonance_mechanism_v2.md: Technical foundation v2.0 formalizing
  Holochain validation lifecycle as resonance (5-property irreducible
  kernel, formal definitions, irreducibility proof)
- flossi0ullk-landscape-recursive-analysis_3-24-26_perplexity_computer_used.md:
  Four-layer recursive landscape research report with epistemic tier
  taxonomy ([TESTED]/[VENDOR-CLAIMED]/[COMMUNITY-REPORTED]/[RESEARCH-ONLY]/
  [SPECULATIVE]) and red-team-of-red-team verification

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
@kalisam kalisam changed the title Update dependencies and formalize HARVEST Protocol specifications ADR batch + substrate bridge + Seam 1 consensus gate + signed gradient Apr 5, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants