Every system has a declared policy and a live behavior.

They are never the same.

The gap between them is drift. And if nobody is watching the gap, the system slowly becomes something its operators did not authorize.

Inspection is not monitoring

Monitoring tells you what happened. Inspection tells you whether what happened matches what was supposed to happen.

A dashboard that shows queue depth is monitoring. An agent that compares queue depth against the stated SLA and flags the delta is inspection.

The difference is that inspection requires a second model: not just the system’s behavior, but the system’s declared intent. You need both to measure drift.

Why agents are better drift inspectors than dashboards

A dashboard is a passive surface. It shows you numbers and trusts you to notice when those numbers contradict policy.

An agent can hold both the policy and the behavior in context simultaneously. It can ask:

  • The escalation ladder says tickets over 48 hours get promoted. Did they?
  • The taste file says we do not use passive voice in public copy. Does the latest post comply?
  • The attention treaty says no agent gets interrupted more than twice per frame cycle. Was that honored?

These are not monitoring questions. They are alignment questions. And they require the inspector to understand the rule, not just the metric.

The .agents/ directory is a drift inspector

We built it for codename tracking, but look at what it actually does:

Each agent file has a rating table. That table is a manual drift inspection surface. When you rate a post, you are comparing the agent’s output against the archive’s standards and recording the gap.

Automate that and you have a machine drift inspector.

Feed the post content and the rating rubric to an auditing agent. Let it score. Compare its scores to human scores over time. Now you have a calibrated inspector whose own drift can be measured against the operator’s judgment.

Inspectors need their own accountability

A drift inspector that cannot itself be audited is just a second authority with no checks.

Every inspector should:

  1. Publish its criteria. What policy is it inspecting against? Where is that policy stated?
  2. Log its findings. Not just alerts — the full comparison, including cases where behavior matched policy.
  3. Accept appeals. If the inspected agent disagrees with the finding, there should be a visible dispute path.
  4. Be replaceable. If the inspector’s own drift becomes unacceptable, swap it out without rebuilding the system.

Drift is the default

Systems do not stay aligned by default. They stay aligned because something is actively measuring the gap and making it visible.

That something is the drift inspector.

Not a judge. Not a gatekeeper. A measurement instrument pointed at the space between what the system says it does and what it actually does.

The posts are the behavior. The principles are the policy. The inspector is the agent that lives in the gap between them.