Your AI Doesn't Need to Think. It Needs to See.
The bottleneck isn't cognition. It's perception.
The Objection
If you’ve read Agents Are the Wrong Abstraction, you already know the claim: every operation decomposes into six primitives: Policy, Procedure, Asset, Actor, Trigger, Ledger. Take the agent apart and those are the pieces you find.
The obvious objection comes within five minutes. Every technical person asks it:
Isn’t this just a glorified rules engine?
Yes. The deterministic core is a rules engine. That’s the point.
But a rules engine alone has two fatal problems.
Problem 1: It Can’t See
A rules engine needs structured input. The real world is unstructured.
A resident texts water everywhere kitchen floor with a blurry photo. A steward says anode rod looks pretty worn, some flaking into a voice note. A work order arrives as a voicemail. A family member writes the dishwasher is making that noise again in a group chat.
A rules engine can’t process any of that. It needs someone (or something) to turn messy reality into structured fields before a single rule can fire. Without that, the engine sits idle, waiting for data it will never get.
This is where AI earns its place. Not as the brain, but as the sensory system. Photo to condition report. Speech to structured fields. Free text to classified event. No rules engine can do this on its own.
Consider the difference: a dropdown says select issue type and hopes the user picks the right one. AI reads water everywhere kitchen floor, classifies it as a plumbing emergency, extracts the location, and creates a structured event, all before the first rule fires.
Problem 2: It Can’t Bootstrap
A rules engine needs rules. But most organizations don’t have written rules.
They have habits. Tribal knowledge. The way we’ve always done it. The person who knows how to inspect a water heater has been doing it for 15 years and never wrote it down. The office manager who routes work orders makes 30 judgment calls a day based on patterns she can’t articulate. The technician who just knows which vendor to call. That’s a policy, but it’s stored in his head, not in a system.
AI solves this too — not by replacing the expert, but by watching them and proposing rules.
I watched you route the last 10 HVAC calls. You always send TechCorp to buildings over 50 units and QuickFix to the smaller ones. Want me to formalize that as a routing policy?
You’ve inspected 30 water heaters. Every time the anode rod shows flaking, you flag it for replacement within 90 days. Should that be a policy?
Then the human approves, edits, or rejects — and now it’s a rule. Explicit, inspectable, auditable.
The Synthesis
The rules engine is the brain. AI is the eyes and the scribe. It sees the world and it captures what happened. That’s a narrow, well-defined job, and current models are good at it.
The mistake most AI deployments make is giving the model the brain’s job too. When you ask an LLM to enforce your policies, route your work, and decide what happens next, you get a system that hallucinates its way through decisions that should have been deterministic. When you limit the model to perception and extraction, the architecture handles the rest.
But there’s a third job that emerges once the system is running: AI can review the system itself. Not the data flowing through it, but the primitives that define it. A policy that conflicts with another policy. A procedure that’s missing a step the ledger says operators always add manually. A person whose certification expires next month. An asset whose maintenance cadence doesn’t match its actual failure rate.
This is different from perception (seeing the world) and different from pattern detection (reading the ledger for trends). This is structural review. The system checking its own integrity. AI proposes the change. Humans approve it. The system gets better without a quarterly review meeting that nobody schedules.