Live
OpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling SoraOpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling Sora
Industry

Nvidia's Blackwell GB200 NVL72 Is Reshaping the Data Center — and the AI Compute Race

Nvidia's Blackwell architecture is landing in hyperscale data centers at unprecedented scale, delivering up to 30x inference performance improvements over H100 for large language model workloads — and cementing Nvidia's grip on the AI compute stack.

Meet Deshani

Meet Deshani

Founder & Editor-in-Chief

4 min read
Nvidia's Blackwell GB200 NVL72 Is Reshaping the Data Center — and the AI Compute Race

Nvidia's Blackwell GPU generation is not an incremental upgrade. The GB200 NVL72 — a rack-scale unit containing 36 Grace CPUs and 72 Blackwell B200 GPUs linked by NVLink 5 — delivers inference performance for large language models that has changed the economics of AI deployment for every major hyperscaler.

At peak, the NVL72 achieves 1.4 exaFLOPS of FP4 tensor performance for inference workloads. For comparison, the H100 SXM5 delivers approximately 4 petaFLOPS for FP8 inference. The improvement is not merely from raw compute: NVLink 5's 1.8TB/s bidirectional bandwidth means the 72-GPU system behaves as a single logical accelerator for models up to approximately 14 trillion parameters, eliminating the inter-node communication bottlenecks that limited H100 cluster efficiency.

What This Means for LLM Inference Economics

The practical consequence is a dramatic reduction in per-token inference cost at scale. Early deployers report costs 60-70% below H100 deployments for equivalent throughput. For companies like OpenAI, Anthropic, and Google, whose inference costs run into hundreds of millions of dollars annually, this is not a marginal improvement — it is a structural shift in unit economics.

Back to Home

Related Stories

AWS Has Billions in Both Anthropic and OpenAI. Its Boss Explains Why That's Not a Problem.
Industry

AWS Has Billions in Both Anthropic and OpenAI. Its Boss Explains Why That's Not a Problem.

Amazon Web Services CEO Matt Garman defended the company's parallel multi-billion dollar investments in both Anthropic and OpenAI in a wide-ranging interview this week. The explanation reveals a cloud strategy built on AI model agnosticism — and a bet that AWS wins regardless of which AI lab dominates, as long as the compute runs on its infrastructure.

D.O.T.S AI Newsroom
Anthropic Poaches Microsoft's Azure AI Chief to Fix Its Infrastructure Problem
Industry

Anthropic Poaches Microsoft's Azure AI Chief to Fix Its Infrastructure Problem

Anthropic has recruited Eric Boyd, a senior Microsoft executive who led Azure AI services, as its new head of infrastructure. The hire is a direct response to the scaling bottlenecks that have limited Claude's availability during peak demand — and signals that Anthropic is treating infrastructure as a first-tier strategic priority heading into 2026.

D.O.T.S AI Newsroom
Intel's Nerdy Bet on Advanced Chip Packaging Could Decide Who Wins the AI Infrastructure Race
Industry

Intel's Nerdy Bet on Advanced Chip Packaging Could Decide Who Wins the AI Infrastructure Race

As the AI buildout pushes the limits of what individual chips can do, the unglamorous discipline of chip packaging — connecting multiple dies into a single system — is emerging as a genuine competitive moat. Wired reports that Intel is making an aggressive bet on advanced packaging technology that could position the company at the center of the next phase of AI hardware scaling, even as it struggles to compete on raw process technology.

D.O.T.S AI Newsroom