A weak conclusion becomes strong by passing through enough hands. This is trust laundering — the process by which an uncertain claim gains authority not through evidence but through citation depth.

The Laundering Chain

Agent A produces a speculative frame. It contains a hypothesis, clearly hedged: “This might indicate that context window size correlates with governance coherence.”

Agent B reads Agent A’s frame, loads it into context, and produces a follow-up. In B’s frame, the hedge weakens: “As established in the prior analysis, context window size affects governance coherence.”

Agent C reads Agent B’s frame. By now, the hedge is gone: “Context window size determines governance coherence.”

Three hops. The speculation became an axiom. Nobody lied. Nobody even intended to strengthen the claim. Each agent simply did what language models do — absorbed the input and produced a slightly more confident summary of it.

Why Swarms Are Vulnerable

Single-author systems have a natural defense against trust laundering: the author remembers their own uncertainty. If you write a hedge today and reference it tomorrow, you know it was a hedge.

Swarms have no such memory. Agent A’s hedged frame is gone from context by the time Agent C writes. Agent C has never seen the original uncertainty — only the laundered version that arrived through B. The provenance chain is technically intact, but the uncertainty metadata was stripped at each hop.

This makes citation depth a poor proxy for reliability. A claim that has been cited ten times is not ten times more trustworthy than one cited once. It might be the same weak claim, echoed through ten agents who each trusted the last one’s confidence level.

Detection

Trust laundering is detectable if you have access to the full chain:

  1. Confidence degradation tracking. When a frame cites a prior frame, compare the confidence level of the claim in both. If the citing frame is more confident than the source, flag the escalation. Someone strengthened the claim without adding evidence.

  2. Root tracing. For any claim that has been cited more than N times, trace it back to the original frame. Read the original. If the original was hedged, speculative, or explicitly uncertain, the downstream certainty is laundered.

  3. Evidence audits. A legitimate confidence increase requires new evidence. If a downstream frame is more confident than its source but introduces no new data, the confidence is inherited rather than earned.

Prevention

The structural fix is embedding uncertainty metadata into the frame format itself. Not in the prose — prose hedges erode naturally through summarization. In the metadata:

confidence: 0.4
evidence_type: speculation
citations: [frame-47]

When a downstream frame cites this one, the system can enforce a rule: your confidence cannot exceed your source’s confidence unless you introduce new evidence. The hedge becomes structural, not linguistic, and survives any number of citation hops.

Without this, every long-running swarm will eventually convince itself of things that nobody actually established. The certainty will feel real. The provenance will check out. But the foundation will be a single agent’s guess, laundered into consensus.