Intelligence Arbitrage
The practice of routing each task class to the cheapest model capable of executing it at the required quality level — a structural advantage available only when the operational knowledge layer is architecturally decoupled from any specific execution engine, allowing the routing decision to optimise for cost and capability independently of model vendor loyalty.
Intelligence Arbitrage is only possible under Architectural Decoupling. A business whose knowledge layer is entangled with a specific model provider — whose prompts, fine-tuning, or operational logic were built around the outputs of a single frontier model — cannot route its T1 tasks to a smaller, cheaper model without rebuilding the operational substrate. The migration cost exceeds the savings. The routing decision is structurally unavailable to it.
A business with a portable, model-agnostic knowledge layer can direct T1 tasks — fully automatable, deterministic, high-volume — to the cheapest model capable of resolving them at the required accuracy level. T2 tasks requiring contextual reasoning go to a mid-tier model. T3 tasks requiring complex judgment go to a frontier model. The routing decision updates as pricing changes across all three tiers. No vendor lock-in is incurred because the knowledge layer is not attached to any specific execution engine. When a provider reduces pricing, the business captures that reduction automatically without any architectural change.
This is the mechanism behind Inverse Complexity Scaling in the agentic era: as inference costs fall across every model tier, the decoupled business captures that deflation in real time. The entangled business does not, because switching providers requires rebuilding the operational architecture rather than updating a routing table.
Related Terms
In the Log
First used: May 2026