Mar 2026

Encoding the Human

Human judgment can be encoded into agentic systems. Not approximated. Not simulated. Encoded. Start with frameworks, not rules. Build learning cycles that write back into the frameworks. Place the human at exactly one point.

Human judgment can be encoded into agentic systems. Not approximated. Not simulated. Encoded. And when you do it properly, the system doesn't just follow instructions. It self-aligns to the human it serves.

This is the idea most people building agents are missing. They treat the model's capability as the ceiling. But capability without judgment is just fast guessing. The thing that makes an output trustworthy, the reason a senior operator would sign off on it, is never the raw intelligence. It's the taste, the contextual awareness, the understanding of "what good looks like here, for this audience, in this situation." That's judgment. And it lives in people.

The question is whether you can get it out of people and into the system in a way that compounds.

You can. Here's how.

Start with frameworks, not rules.

The first step is capturing how your best people think, not just what they decide. Rules encode decisions: "if X, do Y." They're brittle. They break the moment context shifts, and in live environments, context always shifts. Frameworks encode reasoning: what factors matter, what trade-offs exist, what signals carry weight, what this audience cares about versus what that one does.

When an agent operates against a framework, it doesn't follow a script. It reasons against a standard. It weighs factors the way a human expert would, and it attaches a confidence score to every call it makes. High confidence, it proceeds. Low confidence, it surfaces the decision for a human. The system knows what it knows and knows what it doesn't.

This works across any domain. A framework might encode how a senior risk analyst assesses counterparty exposure. How an experienced clinician triages symptom patterns against patient history. How a legal partner distinguishes boilerplate from clauses that need negotiation. How a team lead reads whether a customer needs empathy or a solution. The structure is the same. You're giving the agent the mental model of your best person and letting it apply that model at scale.

Critically, frameworks also define how agents read and interpret data. A number without interpretation rules is dangerous. The framework doesn't just say "use this metric." It says how to weight it, when it's stale, what constitutes an outlier, when a trend is significant versus noise, and what context changes the meaning. The same data point means different things to different audiences in different situations. The framework encodes those reading rules so the agent interprets data the way a senior operator would, not just retrieves it.

Then build learning cycles that write back into the frameworks.

Frameworks give the agent a starting point. But they're human-written, and humans can't anticipate every context. The question is: how do the frameworks get better over time without losing human control?

This is where learning cycles come in. And they need to be designed per framework, not bolted on generically.

When a human reviews an agent's output and edits it directly, the difference between the original and the edited version is a precise signal. What did they change? What did they cut? What did they restructure? That delta is captured and recorded. When a human sends direction back instead, "too aggressive for this audience," "lead with the data, not the narrative," that's a different signal captured alongside it.

But here's the part most people get wrong. These signals don't automatically change the system. A single correction is just an observation. Two corrections are two observations. The system records them, links them to the specific framework and context that produced the output, and waits.

As observations accumulate, the system detects patterns. The same type of correction, for the same type of context, repeated across multiple instances. At a threshold, these patterns become soft runtime adjustments, weighted suggestions the agent factors into its next run. These are automatic but deliberately gentle. They carry a confidence score. They decay if not reinforced. They're nudges, not rewrites.

This is the first speed. The system adapts its behaviour at the edges without touching the core frameworks. It's responsive but conservative.

The second speed is where frameworks actually evolve. When a pattern sustains across enough instances, with enough consistency, the system doesn't just adjust at runtime. It proposes a change to the hard-written framework itself. It surfaces the pattern, the evidence behind it, and a specific recommendation for what should change.

And then a human decides.

The human reviews the proposal, examines the evidence, and either approves the change, modifies it, or rejects it. If approved, the framework is updated. The previous version is preserved for rollback. The audit trail is complete. The system's reasoning structure has evolved, but only because a human examined the evidence and made the call.

This is self-alignment in the truest sense. The system doesn't drift. It doesn't silently mutate its own logic. It surfaces what it's learning from human behaviour and asks the human to decide whether that learning should become a permanent structure. The frameworks stay human-owned. But they improve at the pace of the system's observation, not the pace of the human's memory.

Where the human sits in the loop matters as much as whether they're in it.

Most teams put humans everywhere. Review this step, approve that step, confirm this intermediate output. Every checkpoint is a bottleneck, and none of them produce a clean learning signal because the human is intervening before the agent has finished thinking.

The better architecture places the human at exactly one point: after the agent has done its complete work, before anything consequential happens. The agent reasons, selects, drafts, validates. Then the human sees a finished output with the agent's full reasoning visible. Their response, whether they approve, edit, or redirect, is reacting to a complete decision, not interrupting a partial one.

This single intervention point produces the richest signal. And it's the only place where both feedback paths, direct editing and verbal direction, are meaningful. The human is responding to what the system committed to, not what it was still figuring out.

This is the design pattern for any agent in a live environment.

Compliance agents drafting regulatory assessments. Sales agents qualifying leads and writing outreach. Clinical agents triaging patient intake. Operations agents routing and resolving support cases. Underwriting agents assembling risk profiles. Legal agents reviewing contract language. Content agents producing material that represents a brand.

Every one of these touches real people, real decisions, real consequences. Every one of them requires judgment that the base model doesn't have out of the box. And every one of them can be built with this architecture: frameworks that encode expert reasoning and data interpretation rules, a human at the decision boundary, and learning cycles that surface what the system observes back into the frameworks the human controls.

Any agent that ships into production without this faces one of two fates. Either humans review everything forever, and the system never becomes more than expensive autocomplete. Or humans are removed too early, and the system runs on raw model capability without the contextual judgment that makes outputs trustworthy.

The third path is building systems that self-align to the humans they serve. Systems where every human interaction is observed, where patterns are surfaced with evidence, where framework changes are proposed but human-decided, and where the distinction between soft runtime adjustment and permanent structural change is explicit and enforced.

But self-alignment alone isn't enough. It needs a structure underneath it.

The learning cycles, the frameworks, and the human at the decision boundary. These are the visible architecture. What holds them together is less visible but equally critical: well-defined skill rules that govern how each agent behaves, context engineering that ensures the right information reaches the right agent at the right time, and metadata that gives the system memory and traceability across every decision it makes.

Skill rules define the boundaries of each agent's reasoning. Not what it decides, but how it's allowed to think. What it reads, what it queries, what it passes downstream, what it flags. Without precise skill definitions, agents drift. They hallucinate scope. They make calls they shouldn't be making.

Context engineering determines what an agent sees when it makes a decision. A model with perfect reasoning and wrong context produces a confidently wrong output. Getting context right means being deliberate about what travels with a piece of work through the system, what gets enriched at each step, what gets stripped, and what gets inherited from parent processes. The system needs to digest its inputs properly, extract what matters, and carry that substance forward so every downstream agent is working with full knowledge, not partial signals.

Metadata is the system's memory. Every decision tagged, every adjustment traced, every human correction linked back to the output it corrected and the framework that produced it. Without metadata, the learning cycles have nothing to learn from. With it, the system builds a complete, auditable record of how its judgment evolved and why.

This is the full picture. Self-aligning learning cycles on top. Skill rules, context engineering, and metadata underneath. Together, they create an agentic system that doesn't fail, not because it's infallible, but because it knows what it doesn't know, it learns from every human interaction, and it carries the full context of its reasoning at every step. A system that stays uniquely human, no matter how autonomous it becomes. Because the intelligence it runs on was always human to begin with.

Next article

Agentic Scope

Continue reading the design philosophy.

Read next → or back to all articles