Digi Power X Inks $2.5B Cerebras AI Data Center Deal

Digi Power X signed a deal worth up to $2.5 billion with Cerebras Systems to deliver AI data center capacity, expanding compute infrastructure for the wafer-scale chip pioneer's growing inference business.

Share
Digi Power X Inks $2.5B Cerebras AI Data Center Deal

Digi Power X has signed an agreement worth up to $2.5 billion with Cerebras Systems to deliver AI data center capacity, marking one of the larger infrastructure deals of the year and a significant vote of confidence in Cerebras' wafer-scale compute strategy. The arrangement underscores how the AI boom is reshaping data center economics, with specialized silicon vendors increasingly locking in long-term capacity to support inference workloads at scale.

The Deal in Context

Digi Power X, a digital infrastructure operator, will provide hosting and power capacity for Cerebras' AI compute deployments. While full terms were not disclosed, the headline figure of up to $2.5 billion positions this as a multi-year, multi-site commitment likely tied to Cerebras' aggressive expansion of its inference cloud.

For Cerebras, securing dedicated capacity is strategic. The company has spent the past year positioning its Wafer-Scale Engine (WSE-3) chips as the fastest path to high-throughput inference for large language models, claiming order-of-magnitude speed advantages over GPU-based clusters for certain workloads. But translating those chip-level benchmarks into a viable cloud business requires power, cooling, and physical footprint — exactly what Digi Power X is contracted to deliver.

Why Wafer-Scale Matters for Generative Media

Cerebras' WSE-3 is the largest chip ever built, packing roughly 4 trillion transistors and 900,000 AI-optimized cores onto a single silicon wafer. Unlike GPU clusters, which require sharding models across many chips and incur communication overhead, the WSE keeps weights and activations on-die, dramatically reducing latency for transformer inference.

This architecture is increasingly relevant beyond text generation. Video diffusion models, audio synthesis pipelines, and real-time voice cloning systems are bottlenecked by memory bandwidth and inter-chip communication — precisely the constraints Cerebras targets. As synthetic media workloads shift from offline batch generation toward real-time interactive use cases (live avatars, streaming voice agents, interactive video editing), the case for specialized inference silicon strengthens.

The Infrastructure Arms Race

The Digi Power X deal is part of a broader pattern of multi-billion-dollar AI infrastructure commitments reshaping the industry:

  • OpenAI and Oracle have signed compute agreements reportedly exceeding $300 billion over multiple years.
  • Anthropic has committed tens of billions to AWS Trainium capacity.
  • Nvidia continues to anchor the bulk of frontier-model training, but inference is fragmenting toward specialized providers including Cerebras, Groq, and SambaNova.

Cerebras' approach — partnering with infrastructure operators rather than building its own data centers — mirrors the asset-light strategy used by some hyperscalers' specialized AI divisions. It allows Cerebras to focus capital on silicon and software while transferring the power, real estate, and construction risk to partners like Digi Power X.

Implications for Synthetic Media Builders

For developers working on AI video, deepfake detection, voice cloning, or generative audio, expanding non-Nvidia inference capacity has practical consequences:

Lower inference costs at scale. Competition among inference providers is already pushing per-token pricing down. Cerebras has publicly advertised inference speeds of thousands of tokens per second on Llama-class models, and capacity expansion translates that throughput advantage into broader availability.

Reduced GPU dependency. The synthetic media stack has been overwhelmingly Nvidia-centric, from training Stable Diffusion variants to running real-time face-swap pipelines. Alternative silicon at scale gives builders optionality — particularly for latency-sensitive applications like live deepfake detection or interactive avatar systems.

New deployment patterns. Wafer-scale inference unlocks model architectures that were previously impractical to serve, including very long context windows and large mixture-of-experts models. Expect this to influence the next generation of multimodal models combining text, image, and video understanding.

What to Watch

The deal's actual value will depend on how aggressively Cerebras can fill the contracted capacity with paying inference customers. Deployment timelines, specific site locations, and power sourcing — particularly important given the AI industry's rapidly growing electricity footprint — remain to be disclosed. Watch for follow-on announcements detailing which Cerebras inference cloud regions come online first and whether major generative media platforms sign on as anchor tenants.

The infrastructure layer rarely makes headlines in synthetic media coverage, but every advance in AI video quality, voice cloning realism, and deepfake detection ultimately runs on deals like this one.


Stay informed on AI video and digital authenticity. Follow Skrew AI News.