Token usage — dubbed 'tokenmaxxing' — is emerging as the AI era's equivalent of measuring lines of code: easy to track, easy to game, and disconnected from real productivity. Meta's internal token-consumption leaderboard sparked industry backlash after going public. Engineering leaders share alternative frameworks: a four-level 'cognitive delegation' model inspired by the executive chef analogy (measuring how much mental work engineers offload to AI agents), and a self-reporting approach that works in high-trust organizations. Experts agree token usage is a useful early adoption signal but a poor productivity metric, and that combined metrics — not a single number — are needed. No consensus exists yet on the right replacement, and tooling to connect token spend to shipping outcomes is still immature.
Table of contents
Your inbox, upgraded.The easiest metric to gameFrom line cook to executive chef: a skill-level-based frameworkMore like thisThe opposite approach: self-reporting and trustHard to measure, but it has to be measuredThe tokenmaxxing backlashSort: