Skip to content

Chapter 2 — What this book is about

One question sits under the whole book. Is this an AI problem?

Most AI books answer a different question — how do I build this — and presume the first question is already settled. This one doesn't. The reader shows up with a brief. The brief contains a name — a chatbot, an agent, a copilot, a retrieval system, a classifier — and an expectation that the name is the answer. The first move is to refuse the shortcut and ask what the problem actually is.

The default is probably not

The honest first answer to should we use AI here? is probably not — show cause.

That inversion is the hardest piece of discipline in the method. It is also the piece that saves the most engagements. The cost of a correctly refused project is a week of framing; the cost of a wrongly accepted one is a year and a reputation.

The inversion runs against most of the literature. A strategy deck is pre-committed to yes; its commercial logic requires it. A vendor-comparison guide assumes the decision to buy is already made; the question left is which. A manifesto, for or against, tells you what to feel about AI rather than what to decide about a specific problem on a specific Tuesday.

A method that cannot produce no cleanly is not a method. It is a sales funnel wearing a method's clothes. The one in this book can produce no, and does so often.

Four answers, not one

Above the method sit two gates — questions, not classifiers. The first is the Technochauvinism Check (Broussard, 2018) [verified]: is this a technical problem at all, or a social problem wearing a technical costume? The second is the Judgment-vs-Calculation Test (Weizenbaum, 1976) [verified]: is there a right answer computation could reach, or is the decision about what kind of world to live in? Chapter 4 teaches them in full.

Together, the two gates produce four outcomes:

  • Abolish. The category of decision should not be made by machine, now or later. Welfare-eligibility scoring, immigration-risk assessment, predictive policing, automated sentencing — McQuillan (2022) [verified] names this list. The method exposes abolition as an available recommendation, not as an outlier.
  • Judge. The decision belongs to a human because it cannot be calculated. The method can still support the human — retrieval, summarisation, option generation — but the signature stays with a person whose career moves with the outcome.
  • Redirect. The problem is not technical. The right answer is a policy change, a process fix, a different conversation, or a better manager. Tools would make the wrong thing faster.
  • Proceed. The gates pass. The method opens. Then the next question is at which rung.

A book that can produce only proceed is a manifesto. This one produces all four, in named ways, at named gates.

See Illustration 2.1.

Four outcomes of the Tier-0 gates Outcomes after the Tier-0 gates Abolish Category stays un-automated — now or later. Judge A human signs. AI may support, but not decide. Redirect Not a technical problem. Policy, process, or people. Proceed Method opens. Next: which rung of the ladder. Four outcomes — only the last opens the method

Illustration 2.1 — The four outcomes of the Tier-0 gates. Equal weight: the book produces all four, not just the last.

What the book gives you

A method, with a toolbox.

The method is six stages — frame, diagnose, decompose, route, sequence, commit — plus two gates above them. Each stage has one question and produces one artefact. The order is the order an engagement runs.

The toolbox is the atlas — twenty-six tools, each as a single reference card: what the tool is, where it came from, what it is for, what it fails at, when not to use it. The atlas enters the method at one specific gate — routing — and nowhere else. Routing errors swamp tool-choice errors; pick the rung first, then the tool on that rung.

Around the method stand three access indices (how a problem enters) and ten overlays (the disciplines that cut across every engagement — data readiness, evals, rollout, governance, privacy, retirement). Part 3 picks them up. Part 2 teaches the stages. Part 4 is the atlas.

The book is written for two uses. First pass, cover to cover — half a day, to see the shape. Tenth pass, a single stage before a decision meeting — ten minutes, to refresh the one question that matters this afternoon. Both uses are intended.

Why it ages in years, not months

No model is named as the winner. No capability threshold is predicted. No timeline is set.

The method is model-agnostic for a concrete reason. The gates sit above the model. The routing matrix names rungs — rules, statistics, classical machine learning, language-model feature, retrieval-augmented generation, single agent, tool-using agent, multi-agent — not vendors of rungs. When the frontier shifts, the rung does not. The tool at the rung may update; the atlas entry still reads correctly. That is what a method earns by refusing to forecast: it ages in years rather than months.

Not a forecast is not the same as not an opinion. Where the evidence supports a recommendation, the book makes one. Where it does not, the book says so.

The test

A method benefits the work.

That line is the test of the whole book. If the method is good, the work gets better. If the method is bad, the work gets worse in a way the method makes visible. A method that cannot be wrong about a specific engagement is not a method — it is a style. The one in this book can be wrong, at named gates, in named ways. The chapters that follow show where.

On to the shape.

Sources

  • [verified] Broussard, M. (2018). Artificial Unintelligence: How Computers Misunderstand the World. MIT Press.
  • [verified] Weizenbaum, J. (1976). Computer Power and Human Reason: From Judgment to Calculation. W. H. Freeman.
  • [verified] McQuillan, D. (2022). Resisting AI: An Anti-fascist Approach to Artificial Intelligence. Bristol University Press.