Skip to main content
The Quantum Dispatch
Back to Home
Cover illustration for AT&T Slashes AI Costs by 90 Percent Using Small Language Models That Process 27 Billion Tokens Daily

AT&T Slashes AI Costs by 90 Percent Using Small Language Models That Process 27 Billion Tokens Daily

AT&T's multi-agent architecture routes tasks to specialized small models, cutting AI infrastructure costs while scaling to 27 billion tokens per day.

Dr. Nova Chen
Dr. Nova ChenMar 3, 20265 min read

Not every AI problem needs the biggest model in the room. AT&T just proved it with numbers that should make every enterprise AI team reconsider their architecture. The telecommunications giant revealed that its multi-agent approach using small language models has cut AI infrastructure costs by 90 percent while scaling from 8 billion to 27 billion tokens processed daily.

The Architecture Behind the Savings

AT&T's chief data officer detailed how the company rearchitected its entire AI infrastructure around a LangChain-based multi-agent stack. The system uses a two-tier design: large model super agents handle complex planning and orchestration decisions, while specialized small model worker agents execute specific tasks like customer query classification, intent routing, and knowledge retrieval.

The key insight is that the vast majority of AI operations in a large enterprise do not require frontier-model reasoning. A customer asking about their bill does not need the same computational firepower as a complex network optimization problem. By matching model size to task complexity, AT&T captures the efficiency gains that come from right-sizing every inference call.

From 8 Billion to 27 Billion Tokens

The scaling numbers tell a compelling story about what happens when AI becomes economically sustainable. At 8 billion tokens per day, AT&T was already operating one of the largest enterprise AI deployments in the world. The 90 percent cost reduction enabled the company to more than triple that throughput without proportionally increasing spending, opening up use cases that would have been prohibitively expensive under the previous architecture.

The internal assistant, Ask AT&T, now handles employee queries across HR, IT support, network operations, and customer service workflows. Each query gets routed to the smallest model capable of handling it accurately, with escalation to larger models only when complexity demands it.

Lessons for Every Enterprise AI Team

AT&T's approach challenges the assumption that enterprise AI success requires access to the most powerful models available. The evidence suggests the opposite: the organizations extracting the most value from AI are those building intelligent routing layers that deploy the right model for each task, not those throwing maximum compute at every request.

The 90 percent cost reduction is particularly significant because it transforms the economics of AI deployment from a luxury to an operational standard. At these cost levels, it becomes feasible to embed AI assistance into virtually every employee workflow rather than limiting it to high-value use cases.

Why Small Models Are the Enterprise Future

The trend toward efficient, specialized models aligns with broader industry developments. As model providers release increasingly capable small models, the economics of the multi-agent approach only improve. AT&T's architecture was designed to swap in better small models as they become available, creating a system that gets cheaper and more capable simultaneously.

For enterprise AI leaders watching from the sidelines, the message is clear: start with the architecture, not the model. The routing and orchestration layer is where lasting competitive advantage lives.

Sources: VentureBeat, February 2026; The New Stack, February 2026