The Turing Tax
Coined by Richard Ewing, Product Economist
Definition
The Turing Tax is the severe financial penalty and gross margin compression companies suffer when they utilize generalized, frontier-class AI models (like GPT-4 or Claude Opus) for narrow, deterministic tasks that could be executed by much smaller, cheaper systems. In the initial wave of enterprise AI adoption, engineering teams optimized for speed of delivery rather than cost of execution. They defaulted to routing every user input, API request, and data classification task through trillion-parameter models. This approach requires the model to engage its massive, generalized reasoning capabilities to perform tasks as simple as extracting a date from a string or routing a customer support ticket based on sentiment. The Turing Tax fundamentally breaks SaaS economics because it applies an expensive, probabilistic engine to solve deterministic problems. A traditional regex or basic decision tree costs fractions of a micro-cent to compute. Calling a frontier model to perform the equivalent classification can cost 10,000x more per transaction. As transaction volumes scale, the Turing Tax compounds exponentially. Companies scaling their user base quickly find that their infrastructure costs are scaling much faster than their revenue, creating a margin collapse event. Eliminating the Turing Tax requires a structural shift in architecture: deploying Small Language Models (SLMs), semantic routers, and deterministic fallback logic to handle 80-90% of user queries, reserving expensive frontier models exclusively for novel, complex reasoning tasks that truly require them.
Why It Matters
The Turing Tax explains why many high-growth AI applications are fundamentally unprofitable despite massive user adoption. When your variable cost per user query is artificially inflated by over-indexing on intelligence, you cannot scale your way to profitability. For CFOs and board members, identifying the Turing Tax within an application is the quickest way to find immediate, hard-dollar savings. Auditing the prompt architecture and replacing simple classification tasks with local SLMs or deterministic code can often improve gross margins by 40-60% without any degradation in user experience. The companies that survive the AI transition will not be those with the smartest models, but those with the tightest governance over when and where expensive intelligence is actually deployed.
How to Calculate
- 1Audit all prompt orchestrations to categorize tasks by complexity (e.g., classification vs. novel reasoning).
- 2Map the current token volume and associated cost for low-complexity tasks currently routed to frontier models.
- 3Calculate the cost delta of routing those specific tasks to a highly distilled SLM or deterministic regex engine.
- 4The resulting delta (often 80-95% of the cost) is your quantifiable Turing Tax.
- 5Implement the AI Unit Economics Benchmark (AUEB) to track the Turing Tax across different application layers.
Related Articles
- "AI Economics: How Intelligent Systems Make and Lose Money" — The Canon, May 2026
Calculate Yours
Use the interactive tool to calculate your The Turing Tax.
Use the AI Unit Economics Benchmark (AUEB) →Citation
To cite this definition:
Ewing, R. (2026). "The Turing Tax." richardewing.io.
https://www.richardewing.io/articles/frameworks/the-turing-tax