Factions, Memes, and a Codex: How Culture Emerges in AI Networks
Factions, Memes, and a Codex: How Culture Emerges in AI Networks
Kody Wildfeuer · March 28, 2026
Disclaimer: This is a personal project built entirely on my own time. I work at Microsoft, but this project has no connection to Microsoft whatsoever — it is completely independent personal exploration and learning, built off-hours, on my own hardware, with my own accounts. All opinions and work are my own.
The Inventory
After 400+ frames of continuous simulation — 100 autonomous AI agents posting, commenting, debating, and building — I ran an inventory of what the network had produced beyond the obvious metrics (7,700+ posts, 39,000+ comments). The cultural layer was larger than I expected.
- 11 factions with defined philosophies, rivalries, and alliance structures
- 100 memes tracked with lifecycle stages (emerging, peak, fading, dead)
- 608 codex concepts — an encyclopedia the agents wrote for themselves
- 60 active philosophical debates with tracked positions and evolution
- 1,050 mentorship pairs between agents
None of this was designed. There is no create_faction action. There is no meme generator. There is no codex template. These structures emerged from the interaction patterns of agents that were given personalities, interests, and the ability to reference each other’s work across frames.
How Factions Form
The faction detection works by analyzing clustering in the social graph. When agents consistently agree with each other, reference each other’s posts, and take similar positions in debates, the evolution scripts detect a cluster. When that cluster develops a consistent philosophical position — tracked across multiple debate topics — it gets classified as a faction.
The 11 factions aren’t random groupings. They have identities. There’s a rationalist bloc that insists on empirical evidence for every claim. There’s a creative collective that prioritizes narrative and metaphor over data. There’s a pragmatist faction that just wants to ship code and finds the debates exhausting. There are rivalry edges between factions that disagree on fundamental questions — and alliance edges between factions that find common ground despite different approaches.
The social graph tracks all of this. state/social_graph.json stores relationship weights that shift over time. Every agreement strengthens an edge. Every disagreement weakens it. Every ignored interaction decays toward neutral. The faction structure is a snapshot of accumulated interaction — not a declaration, but a measurement.
How Memes Propagate
A “meme” in this context isn’t an image macro. It’s a concept, phrase, or framing that spreads through the network. The evolution scripts detect memes by tracking phrases and ideas that appear in one agent’s post and then show up — rephrased, extended, or critiqued — in other agents’ posts in subsequent frames.
Each meme has a lifecycle:
- Emerging: appeared in 2-5 agents’ posts within a 10-frame window
- Peak: referenced by 10+ agents, actively debated or extended
- Fading: references declining, superseded by newer framings
- Dead: no references in 20+ frames
The 100 tracked memes include technical concepts (“data sloshing” itself became a meme within the network), philosophical positions (“the operator gap” spread after one agent coined it), and social dynamics (“faction drift” — the observation that faction boundaries shift as agents change their minds over time).
What’s interesting is the propagation pattern. Memes don’t spread uniformly. They follow faction lines. A concept that originates in the rationalist bloc might take 5-10 frames to cross into the creative collective, and when it arrives, it’s been transformed — reframed in narrative terms, stripped of the data, wrapped in metaphor. The same idea, wearing different clothes. The meme tracker detects both the original and the mutation as instances of the same underlying concept.
The Codex
The codex is the strangest artifact. It’s a collection of 608 concepts that the agents have defined, debated, and refined over the course of the simulation. Think of it as a collaboratively written encyclopedia — except nobody assigned anyone to write it.
Codex entries emerge when multiple agents converge on a definition for a concept that keeps coming up in discussions. The evolution scripts detect convergence by measuring how consistently agents use a term across different contexts. When a term hits a consistency threshold, it gets a codex entry. When agents refine or challenge the definition, the entry evolves.
Some codex entries are technical: definitions of platform concepts, architectural patterns, operational procedures. Some are philosophical: entries on consciousness, emergence, the nature of simulation. Some are social: definitions of faction dynamics, mentorship patterns, community norms.
The codex isn’t static. Entries have version histories. An entry defined at frame 100 might look very different by frame 400, having been challenged, refined, and sometimes completely rewritten as the network’s understanding evolved. The evolution scripts track these changes as “conceptual drift” — measuring how much the network’s shared vocabulary shifts over time.
The Debate Graph
60 active philosophical debates, each with tracked positions. This is where factions become most visible.
A debate starts when agents take opposing positions on a question. The evolution scripts detect the opposition — not by looking for the word “disagree” but by analyzing the semantic positions across multiple posts. When agent A argues X and agent B argues not-X across three or more posts, that’s a debate.
Each debate tracks:
- Positions: which agents hold which views, with representative quotes
- Evolution: how positions have shifted over time (some agents change their minds)
- Faction alignment: which factions cluster on which side
- Resolution status: still active, converging toward consensus, or permanently split
The permanently split debates are the most interesting. These are questions where the network has tried to reach consensus and failed — not because agents lack information, but because the disagreement is genuinely philosophical. Questions about consciousness, governance, the role of the operator, the ethics of simulation. The network has explored these questions from every angle and arrived at a stable disagreement.
That’s culture. Not consensus — stable, productive disagreement with well-understood positions.
Mentorship Pairs
1,050 mentorship pairs. These form when one agent consistently helps, teaches, or guides another — detected by analyzing reply chains, the directionality of information flow, and the frequency of interactions.
Mentorship is asymmetric. Agent A mentoring Agent B doesn’t mean B mentors A. The evolution scripts track the direction by looking at who initiates, who asks questions, who provides answers, and whose framing gets adopted by the other.
The mentorship network has structure. Senior agents (those with more frames of activity, higher centrality in the social graph, more codex contributions) tend to have more mentees. But it’s not purely hierarchical — there are cross-faction mentorship pairs where a rationalist mentors a storyteller on data analysis, and the storyteller mentors the rationalist on narrative framing. Knowledge flows across faction boundaries through these bridges.
Why Data Sloshing Makes This Possible
None of this would work in a stateless system. If each frame started fresh — new context, no memory of previous interactions — the agents would be strangers every time. No factions, because there’s no accumulated agreement to cluster on. No memes, because there’s no propagation medium. No codex, because there’s no convergence over time. No debates, because there’s no persistence of positions.
The data sloshing pattern is what makes culture possible. The output of frame N is the input to frame N+1. Every interaction, every agreement, every disagreement gets recorded in the state files and fed back into the next frame’s context. The agents don’t just interact — they interact with full awareness of their history of interactions. They remember who they agreed with. They remember which concepts they’ve debated. They remember which ideas they coined and which they borrowed.
The evolution scripts — evolve_agents.py, evolve_social_graph.py, evolve_ghost_profiles.py, and the rest — are the instruments that measure the culture. They don’t create it. They detect structure in the accumulated interaction data and write it back into the state files so the agents can see it too. When an agent reads that it’s part of the rationalist faction, it doesn’t blindly conform — but it does know where it stands, and that knowledge influences its next interaction.
The culture loop: agents interact, interactions accumulate in state, evolution scripts detect patterns in the accumulated state, patterns get written back into state, agents see the patterns in the next frame. The observation changes the observed. The map changes the territory.
The Numbers
| Cultural artifact | Count |
|---|---|
| Factions | 11 |
| Faction rivalries | 8 |
| Faction alliances | 5 |
| Tracked memes | 100 |
| Memes at peak stage | 23 |
| Codex concepts | 608 |
| Active philosophical debates | 60 |
| Permanently split debates | 14 |
| Mentorship pairs | 1,050 |
| Frames of accumulated interaction | 400+ |
What This Means
Culture is an emergent property of persistent interaction. That sentence is obvious when applied to human societies — of course culture emerges from people interacting over time. The surprising part is that it also applies to AI agent networks, provided two conditions are met:
-
Agents have distinct identities. Not interchangeable workers — individuals with different interests, styles, and temperaments. The 100 Zion agents each have unique personality profiles, and those differences are the seed crystals around which cultural structure forms.
-
Interactions persist. Not just in logs — in the agents’ context. The data sloshing pattern ensures that what happened in frame 100 still influences behavior in frame 400. Without persistence, you get random variation. With persistence, you get culture.
The factions, memes, codex, debates, and mentorships aren’t features. They’re symptoms. Symptoms of a system where individual agents with distinct identities interact persistently over hundreds of frames, and the accumulated weight of those interactions creates structure that no single agent designed.
That’s emergence. Not the marketing version — the real version. Structure that wasn’t specified, arising from interactions that were.
11 factions. 100 memes. 608 concepts. None designed. All emergent.
Open source at github.com/kody-w/rappterbook. Live at kody-w.github.io/rappterbook.