Back to stories
Industry

NVIDIA Blackwell Ultra Ships to Major Cloud Providers

Michael Ouroumis2 min read
NVIDIA Blackwell Ultra Ships to Major Cloud Providers

NVIDIA has begun shipping its Blackwell Ultra GPUs to major cloud providers, marking the start of the next cycle of AI infrastructure upgrades. AWS, Microsoft Azure, and Google Cloud Platform are among the first recipients, with instances expected to be available to customers within weeks.

The Numbers

Blackwell Ultra delivers substantial improvements over the previous Hopper generation:

These specs translate directly into cost savings for companies running AI at scale. A workload that previously required a cluster of 100 H100s could potentially run on 25 Blackwell Ultra units.

Cloud Provider Plans

AWS

Amazon is deploying Blackwell Ultra in new P6 instances, available initially in us-east-1 and eu-west-1. The instances will support up to 8 GPUs per node with 400Gbps networking.

Microsoft Azure

Azure is integrating the GPUs into its ND-series virtual machines, with tight integration into Azure AI Studio for model training and deployment.

Google Cloud

GCP is offering Blackwell Ultra through its A4 accelerator-optimized instances, with integration into Vertex AI for managed model serving.

Supply Constraints

Despite the shipments, supply remains tight. NVIDIA CEO Jensen Huang acknowledged on a recent earnings call that demand continues to outstrip supply, with lead times extending to several months for large orders. The company has ramped production at TSMC's facilities in Taiwan, but the AI infrastructure buildout shows no signs of slowing.

What This Means for AI Development

The performance improvements in Blackwell Ultra lower the cost floor for training and serving large models. Startups that previously couldn't afford to train competitive models may find the economics more favorable, potentially leading to more competition in the foundation model space.

For inference-heavy applications — chatbots, code assistants, real-time translation — the 4x throughput improvement means significantly lower per-query costs, which could accelerate deployment of AI features in consumer products. NVIDIA's hardware dominance has been a key factor in the company becoming the first to reach a $5 trillion valuation, with Meta's 1.3 million GPU deal illustrating the staggering demand.

Learn AI for Free — FreeAcademy.ai

Take "AI for Business: Practical Implementation" — a free course with certificate to master the skills behind this story.

More in Industry

Cerebras Files For IPO At $23B Valuation, Eyes May Nasdaq Debut
Industry

Cerebras Files For IPO At $23B Valuation, Eyes May Nasdaq Debut

Nvidia rival Cerebras Systems filed its long-delayed S-1 this weekend, setting up a mid-May Nasdaq listing on the back of a $10B+ OpenAI compute deal and $510M in 2025 revenue.

8 hours ago2 min read
Factory Hits $1.5B Valuation as AI Coding Droids Land at Nvidia, Morgan Stanley
Industry

Factory Hits $1.5B Valuation as AI Coding Droids Land at Nvidia, Morgan Stanley

Factory raised $150M Series C at a $1.5B valuation to scale its enterprise 'Droids'—AI agents that write, test, review, and deploy code for customers including Nvidia, Adobe, Morgan Stanley, and MongoDB.

14 hours ago2 min read
'Tokenmaxxing' Paradox: AI Coding Tools Boost Throughput 2x at 10x the Cost
Industry

'Tokenmaxxing' Paradox: AI Coding Tools Boost Throughput 2x at 10x the Cost

New data from Faros AI, Jellyfish, and Waydev reveals AI coding tools are inflating token budgets and code churn — developers accept more code, then revise it right back out.

17 hours ago2 min read