← back

The Grown-Up Rules for Plugging Cognition Into Everything

preamble

We used to wire electricity into everything. Now we’re wiring cognition into everything.

Agents, copilots, automated workflows – whatever you call them – are already touching code, money, customers, infrastructure, and law. We’re letting semi-autonomous systems act on our behalf in places where mistakes are very expensive.

Most of the industry is still acting like this is a toy.

This doc is the minimum bar for adulthood.

If you want to plug cognition into your company, you don’t get vibes and hype. You get constraints, logs, roles, and responsibility.

These are the minimum rules to be a grown-up about agents. If you’re below this bar, admit you’re still playing with matches.

rule 1 — no unlogged cognition

No ghost workers in core systems.

If an agent can:

Then every action must be logged:

No log = no trust.

If we can’t replay what your “smart system” did, it doesn’t belong anywhere near critical workflows.

rule 2 — humans own outcomes

Responsibility is not automatable.

Every agentic system has a named human steward.

That person:

There is no “the AI did it.” The stack ends on a human.

rule 3 — capability follows competence

No L1 brain gets L5 powers.

You don’t hand a med student solo brain surgery. You don’t hand a brand-new agent root access.

Define clear capability levels, e.g.:

To move a system up a level, it has to pass evals on:

No more “well, the demo looked good so we gave it prod access.”

rule 4 — sandbox before surface area

Everything dangerous grows up in a box first.

Deployment should look like:

  1. Sandbox phase

    • Synthetic or restricted data
    • Heavy logging
    • Adversarial tests
  2. Shadow phase

    • Runs alongside humans on real data
    • Proposes actions, doesn’t execute
    • Human decisions + overrides are logged as training signal
  3. Gradual rollout

    • Start with narrow scope / low blast radius
    • Feature flags, canaries
    • Instant rollback paths

If you “don’t have time” for this, you don’t have time for agents.

rule 5 — tight feedback or no autonomy

If we can’t tell when it’s going off the rails, it doesn’t run unattended.

Autonomy is not a yes/no; it’s bounded by how fast you can detect and recover from bad behavior.

For any unattended agent:

No monitoring = no autonomy. “Set and forget” cognition is how you get surprise, expensive art projects in prod.

rule 6 — interpretability over vibes

We wrap black boxes in white-box workflows.

You will not get perfect “explainability” from LLMs. That’s fine.

For high-stakes use (money, safety, infra, compliance, brand), you still owe:

We don’t need to peek inside the model’s weights. We do need to understand the container we put it in.

rule 7 — socio-technical, not tech-only

Culture is part of the runtime.

Agents don’t live “in isolation.” They live inside orgs with incentives, politics, and human failure modes.

Grown-up deployments answer:

There is no tooling that can save an org whose culture punishes people for pulling the brake. You can’t tech your way out of cowardly leadership.

rule 8 — minimal necessary agency (principle of least agency)

Don’t summon a demigod when you just needed a calculator.

Just because you can jam a general agent into something doesn’t mean you should.

For every use case, ask:

We default to minimal necessary agency. Generality is an escalation, not a starting point.

rule 9 — fail loudly, recover gracefully

No silent corruption.

Bad things will happen. The question is whether you:

Grown-up behavior:

Then you update:

Every agent-related failure becomes training data for both humans and systems.

rule 10 — multi-horizon ethics

Optimize today without poisoning tomorrow.

You’re not just responsible for the next quarter.

Grown-up deployments consider:

You don’t have to be a saint. You do have to stop trading systemic integrity for a little line-go-up dopamine.

If you can’t meet these rules, say so out loud.

Just don’t pretend you’re doing “responsible AI.”

new essays in your inbox.