AI NEWS HUBbyEIGENVECTOREigenvector

The AI industry loves token inflation. Your company shouldn’t

Fast Company Techby Enrique DansMarch 31, 20268 min read0 views
Source Quiz

The AI industry has a quiet addiction problem: It is addicted to tokens. Every new generation of agentic AI seems to assume that the answer to complexity is to throw more context at the model, keep longer histories, spawn more calls, loop over more tools, and let the token meter run wild. The rise of agentic systems, and now projects like OpenClaw , makes that temptation even stronger. Once you give models more autonomy, they do not just consume tokens to answer questions. They consume them to plan, reflect, retry, summarize, call tools, inspect outputs, and keep themselves on track. OpenClaw itself describes the product as an “agent-native” gateway with sessions, memory, tool use, and multi-agent routing across messaging platforms—which tells you exactly where this is going: more autonomy

The AI industry has a quiet addiction problem: It is addicted to tokens.

Every new generation of agentic AI seems to assume that the answer to complexity is to throw more context at the model, keep longer histories, spawn more calls, loop over more tools, and let the token meter run wild.

The rise of agentic systems, and now projects like OpenClaw, makes that temptation even stronger. Once you give models more autonomy, they do not just consume tokens to answer questions. They consume them to plan, reflect, retry, summarize, call tools, inspect outputs, and keep themselves on track. OpenClaw itself describes the product as an “agent-native” gateway with sessions, memory, tool use, and multi-agent routing across messaging platforms—which tells you exactly where this is going: more autonomy, more orchestration and, unless someone intervenes, a lot more token burn.

That trajectory delights almost everyone selling the infrastructure. If billing is based on tokens, more token consumption looks like growth. If you sell the compute behind those tokens, it looks even better. Google said in its October 2025 earnings call that it was processing more than 1.3 quadrillion monthly tokens across its surfaces, or more than 20 times the volume of a year earlier. Nvidia, for its part, has been leaning hard into the economics of inference and agentic AI, highlighting both the demand surge and the opportunity to sell ever more infrastructure into it.

But companies buying AI should look at this very differently. From the customer’s perspective, explosive token growth is not necessarily a sign of intelligence. In many cases, it is a sign of inefficiency.

More tokens are not the same thing as more intelligence

The current industry narrative often treats token consumption as if it were a proxy for progress. Bigger context windows, more reasoning traces, more agent loops, more memory, more retrieval, more interactions. It all sounds impressive.

Explore Topics

  • Artificial Intelligence
  • token
Was this article helpful?

Sign in to highlight and annotate this article

AI
Ask AI about this article
Powered by Eigenvector · full article context loaded
Ready

Conversation starters

Ask anything about this article…

Daily AI Digest

Get the top 5 AI stories delivered to your inbox every morning.

Knowledge Map

Knowledge Map
TopicsEntitiesSource
The AI indu…modellanguage mo…availableproductplatformcompanyFast Compan…

Connected Articles — Knowledge Graph

This article is connected to other articles through shared AI topics and tags.

Knowledge Graph100 articles · 136 connections
Scroll to zoom · drag to pan · click to open

Discussion

Sign in to join the discussion

No comments yet — be the first to share your thoughts!

More in Products