Posts
RSS feed- Rules Decay, Hooks Don't
The difference between writing down a rule and making the system enforce it — illustrated by a 15-line hook.
- Agentic Engineering: Why Less Is More
Tool enthusiasm is often net-negative. Context pollution degrades performance faster than features improve it. The principles that actually work.
- I Made the AI Remind Me of My Own Blind Spots
I kept missing things at the end of AI sessions. So I stopped relying on willpower and systematised the nudge instead.
- CLIs Enforce Structure and Save Tokens — Not Just Discipline
The instinct to add a rule to a skill file is usually the wrong abstraction. A CLI wrapper enforces at the tool level: zero deliberation, zero token cost.
- Three AI Governance Blind Spots No Framework Covers
Most AI governance frameworks are technically-focused risk checklists. Three structural risks are missing from almost all of them.
- AI Evals: Why Teams Build Metrics Before They've Read a Trace
Most teams build evaluators before reading a single trace. The sequence that actually works is the opposite: observe, categorise, then measure.
- AI Vendor Selection Is Now a Values Decision
OpenAI took the Pentagon contract Anthropic refused. Your AI vendor just became a political statement — and enterprise procurement hasn't caught up.
- Software Engineering Principles for AI Instruction Files
LLM instruction files are code. They have the same failure modes — with one interesting twist that changes everything.
- Backtest vs Operational Validation: The Control You Think You Have
A model control that's never fired in production isn't a control — it's a hypothesis. The gap between backtest and operational validation is invisible until someone asks.
- AI Succeeds, Economy Breaks: The Displacement Loop Nobody Models
The standard AI economic models assume wage effects and retraining timelines. They don't model the feedback loop where successful AI deployment reduces the customer base that purchases AI-enabled products.
- The Upstream Constraint Pattern
In digital transformation, the bottleneck is almost always upstream of where the pain is felt. Mox is the cleanest case study.
- Why AI Assistants Make Us Dumber (And What Governance Should Do About It)
The cognitive offloading problem is real. The governance response mostly isn't. There's a specific mechanism at work, and it has a specific fix.
- The Kutta Condition of AI: Engineering Ships Before Theory Catches Up
Aeronautics flew for decades before anyone could explain why wings worked. AI is in the same position. The engineering is ahead of the theory.
- The Failure Mode of AI Advice Isn't Hallucination
The failure mode of AI advice isn't hallucination. It's that it agrees with you. Here's the architecture that fixes it.
- Agent-First CLI Design: TTY Detection as Philosophy
The primary user of my CLI tools isn't me anymore. Designing for that changes everything about how output should work.
- Building My Own Consulting Toolkit Before Day One
Most consultants arrive at a new firm and learn their tools from colleagues. I tried something different.
- Three Crates Before Lunch
I published three Rust CLI tools to crates.io before noon — none existed at breakfast. The interesting part isn't the speed. It's that the bottleneck moved.
- Taste Requires Stakes
AI can simulate aesthetic judgment with impressive fluency. What it cannot simulate is the consequence of being wrong.
- Banks Have an AX Problem They Don't Know About Yet
Banks are building AI agents to call their APIs. Those APIs weren't designed for agent callers. The mismatch is subtle, consequential, and almost nobody is talking about it.
- When to Build vs. When to Wait: The Recurrence Rule for AI Tooling
Most AI tooling debates are actually recurrence debates. The question isn't whether to build — it's how many times you'll need it.
- What Surprised Me Studying for the GARP Responsible AI in Finance Exam
I expected the hard parts to be the technical sections. They weren't. The governance sections were harder, and more useful.
- The AGI Question Nobody Is Asking Correctly
Sequoia says AGI is here. Dan Shipper says we're not there yet. They're both right — they're measuring different things. The question that actually matters is Sholto Douglas's "nines of reliability."
- Three Things AML AI Models Still Get Wrong in 2026
The models aren't the problem. The operating models are. Three structural failures in AML AI from years building these systems inside a bank.
- The AI Job Title Illusion
Two job ads. Same bank. Same week. Same title pattern. Completely different jobs. The AI hiring market has a labelling problem.
- Five Archetypes of AI-Era Business Defensibility
When AI models commoditize, the moat isn't the model. It's the infrastructure AI must flow through but can't replace. Five archetypes of what that looks like.
- Per-Token Pricing Is the 'Megapixels' of AI
We're optimising for the wrong number — and the history of consumer electronics suggests we'll figure this out eventually.
- Skills as Behavioral Nudges: The Lightweight Alternative to Fine-Tuning
We fine-tune models with gradient descent. We nudge agents with skill files. Same goal, radically different cost.
- The Contract Pattern: Hard Gates for AI Agents
AI agents know how to start a task. They don't always know when to stop. The contract pattern is the architectural fix.
- The Real Reason Mox Won (and What It Means for AI Transformation)
Mox didn't win because they hired better designers. They won because they had no legacy to fight. The pattern applies directly to AI transformation.
- AI Governance Category Error: Routing vs. Compliance
Your AI governance framework is a routing spreadsheet pretending to be a compliance programme. Regulators will spot the difference.