Skip to main content
EMil Wu

#17

Golden Circle 1: Theory vs Methodology — What Are You Actually Giving Your Agent?

Mindset 8 min read
A three-tier staircase from theory to methodology to method A three-tier staircase from theory to methodology to method
Theory explains why, methodology guides which approach to pick, method is the concrete steps

In the previous article we talked about an Agent’s timezone blind spot [3] — a 25-hour error that compounded over several days without the Agent ever noticing. I ended that piece by saying: if you really want an Agent to operate autonomously, what you need isn’t a more careful Agent — it’s a methodology.

But “methodology” has become one of those buzzwords that gets thrown around like “framework.” Slap the word “methodology” onto anything and it sounds impressive, but when you open it up, it’s just an SOP.

What these articles are about, though, isn’t how to write SOPs. It’s a complete growth chain: from theory to methodology, from methodology to playbook, from playbook to execution plan. Each layer down is a transition from “general” to “specific,” from “possible” to “certain” — the lower you go, the less flexibility but the higher the executability.

This chain matters because most people, when working with an Agent, habitually jump straight to the bottom layer, handing the Agent a pile of instructions and saying “do it!” The Agent follows through, but the moment it hits a situation the instructions don’t cover, things fall apart — it doesn’t know why it should act a certain way, so it naturally makes bad calls in new situations, or more accurately, it guesses wrong.

So how do you build this chain? Let’s walk through a real case and see how it came together step by step.


What’s the Actual Difference Between Theory and Methodology?

Before the case study, let’s untangle a few concepts that constantly get mixed up.

Theory answers “why.” It’s a systematic set of statements that explain phenomena [1] — organizing the relationships between observations using concepts and logic, telling you why the world works the way it does. But it won’t tell you specifically what to do. Knowing about gravity doesn’t mean you can build a rocket.

Methodology answers “how to approach it and why this approach.” It provides principles and rationale for choosing tools, processes, and steps — a framework that guides action. It’s closer to the ground than theory, but it’s not the ground itself. Agile tells you how to organize software development and iteration cycles, but it won’t write your Sprint Backlog for you.

Method is the concrete steps and tools — what you actually read and follow when you get your hands dirty.

So the hierarchy is:

Theory → Methodology → Method → Execution [2]

Theory explains why, methodology guides how, method is the concrete steps, and execution is doing the work. If you want to “ship right now,” what you need is method-level stuff, because methodology only helps you choose methods, and theory only helps you understand why a method works.

graph TD T["Theory
Explains why"] --> M["Methodology
Guides which approach"] M --> MT["Method
Concrete steps"] MT --> E["Execution
Hands-on work"]
Four-layer architecture: the lower you go, the more concrete — but the less capable of handling new situations

Why bother distinguishing these three layers? Because most of what people give Agents falls into one of two categories: either theory-level (too abstract — the Agent doesn’t know how to execute) or method-level (too specific — breaks down in new situations). What’s missing is the methodology in between, and methodology is exactly what enables an Agent to make reasonable judgments when facing situations it’s never seen before.


A Monitoring System That Reported “All Clean”

A while back, I was assigned a task at work: build a monitoring system for an internal AI Agent platform.

This platform had many different Agents, each handling different responsibilities, all sharing the same Source of Truth (SOT). So we built a monitoring tool to check the platform’s security and governance compliance. After the first round of building and tuning, we had 22 checks running on a daily schedule, and every day the report came back: All Clean.

Sounds perfect, right? Remember what we’ve said before? When an Agent is quiet and everything looks perfect, that’s when you should worry most.

So I used a separate Agent to audit it and discovered the platform had 51 known issues spanning 14 categories. Those 22 checks were only catching a small fraction — the actual detection rate was around 22%. In other words, nearly 80% of the problems were completely invisible to the monitoring system.

Loading chart…
22 Checks vs 51 Known Issues: The Truth Behind All Clean

Worse still, some categories of issues (like cross-Agent data write conflicts, SQL injection risks, and transaction atomicity problems) didn’t even have corresponding checks in the monitoring system. It wasn’t that the checks were poorly written — there were simply no checks designed for those categories at all.

That’s the truth behind “All Clean”: it’s not that there are no problems — it’s that your checks can only see what you designed them to see.


The Agent’s First Instinct: Hand You a Research Plan

When I handed this problem to the Agent and asked it to analyze and propose improvements, it did something perfectly normal. After a round of “thinking,” it gave me a four-step research plan: first inventory all known issues (Ground Truth), then run a full validation pass, then classify the detection gaps, and finally define improvement targets.

At the method level, this plan was fine — clear steps, fully executable. But when I looked at it, I spotted a fundamental problem:

Your research plan is built on coverage rate. Its upper bound comes from “the total number of known issue categories.” Once that number is capped, the Agent stops growing.

This is the ceiling problem — your improvement framework is constrained by its own target. Whether you push coverage from 22% to 80% or even 100%, once all known issues are detected, the system stalls. But in the real world, new problems don’t stop appearing just because your monitoring is thorough.

So I told the Agent:

We also need a corresponding mechanism — a separate loop to explore and discover new problems. This research plan should enable the entire system to achieve a self-growing closed loop, with continuous output, continuous correction, and continuous self-validation.

The Agent accepted this direction and started formalizing the approach, but before it continued, I made a critical decision: don’t rush into execution — first organize this thinking into a methodology.

“Your direction is good, but I want to first organize it into a closed-loop methodology. This methodology should be multi-layered and multi-dimensional. We’ll use it as the summary framework, then continue developing the closed loops from there.”

This “stop and organize” moment is the pivot from the method level back to the methodology level, because if we had charged straight into executing the Agent’s research plan, we’d probably hit the ceiling in three months and have to start all over again.


Three Interlocking Loops — The Architecture That Breaks Through the Ceiling

The Agent then formalized my direction into three interlocking feedback loops:

Detection Loop: Improves the quality and coverage of checks for known issues. This is what most people would do. Despite its name, it actually handles both detection and remediation in a closed loop.

Discovery Loop: Proactively explores unknown problems — running mutation tests on known issues, tracing root causes, making cross-domain inferences — so the scope of known issues keeps expanding and breaks through the ceiling.

Validation Loop: Verifies that the monitoring system itself has no bugs, because a buggy monitoring system is more dangerous than having no monitoring at all — it gives you a false sense of security.

The three loops are “interlocking” because each loop’s output is fuel for the next, and none can modify its own baseline.

  • The Detection Loop produces new checks, which become targets for the Discovery Loop’s mutation testing and regression test cases for the Validation Loop.
  • The Discovery Loop produces newly identified issues, which become improvement targets for the Detection Loop and test instruments for the Validation Loop.
  • The Validation Loop finds bugs in the monitoring system itself, which become fix items for the Detection Loop and blind-spot exploration seeds for the Discovery Loop.

The three loops feed each other, so the system keeps moving forward rather than stalling after one cycle.

graph LR D["Detection Loop"] DC["Discovery Loop"] V["Validation Loop"] D -->|New checks → mutation test targets| DC D -->|New checks → regression test cases| V DC -->|New issues → improvement targets| D DC -->|New issues → test instruments| V V -->|System bugs → fix items| D V -->|Blind spots → exploration seeds| DC
Three interlocking loops: each loop produces two fuel streams feeding the other two, but cannot modify its own baseline

Interestingly, the core logic of the Detection Loop directly borrows from a well-known existing framework: Karpathy’s autoresearch [4]. Its core concept is simple — modify code → evaluate with fixed metrics → if metrics improve, keep it (git commit), if not, discard (git reset) → next iteration. We just swap “modify code” for “modify check logic” and “metrics” for “detection rate,” and we can reuse this iteration framework directly, saving the cost of designing from scratch (we didn’t adopt it wholesale because the metric evaluation approach differs, and modifying it would lose its automatic iteration advantage).

But the most important takeaway from this entire section isn’t the theory — it’s that “stop and organize” moment. If we had charged straight into executing the Agent’s initial improvement plan, we’d probably hit the ceiling in three months and have to start all over.

By this point, the foundation of the methodology is largely in place, but I wanted to make sure it had no holes. So in the next article, we’ll look at what the three cores of this methodology are, what structural vulnerabilities it has, and how to address them on the way to turning it into a Playbook the Agent can act on.


References:

[1] Method vs Theory: Meaning And Differences — theory vs method distinction https://thecontentauthority.com/blog/method-vs-theory

[2] Methodology, Method, and Theory — Helen Kara — three-tier hierarchy https://helenkara.com/2018/02/15/methodology-method-and-theory/

[3] Tips 4: Agent’s Timezone Blind Spot — case from the previous article /en/articles/16-tips-agent-timezone-blind-spot

[4] Karpathy autoresearch — autonomous ML research framework https://github.com/karpathy/autoresearch

Support This Series

If these articles have been helpful, consider buying me a coffee