AI News news

NVIDIA Announces Blackwell Ultra - Next Generation AI Chips Arrive

NVIDIA GPU AI Hardware Blackwell AI News

NVIDIA Announces Blackwell Ultra - Next Generation AI Chips Arrive

NVIDIA has announced the Blackwell Ultra series, the next generation of AI accelerators that deliver four times the performance of the previous Blackwell generation. The new chips are designed to meet the surging demand for AI compute power.

Performance Breakthroughs

Blackwell Ultra B300

  • 20 petaFLOPS FP4 performance
  • 2x memory bandwidth vs B200
  • 2.5TB HBM4 memory
  • 50% better power efficiency

Blackwell Ultra B300X

  • Optimized for inference
  • Lower latency design
  • Better cost-per-query
  • Scalable architecture

GB300 NVL72

  • 72-GPU NVLink system
  • 1.4 exaFLOPS total performance
  • 180TB aggregate memory
  • Single-system training capability

Key Innovations

HBM4 Memory

  • 50% faster than HBM3e
  • Better power efficiency
  • Higher density packaging
  • Improved reliability

Enhanced NVLink

  • 1.8TB/s interconnect speed
  • Unified memory across GPUs
  • Seamless scaling
  • Reduced latency

Transformer Engine 3.0

  • Native FP4 support
  • 2x throughput vs FP8
  • Automatic precision management
  • Training stability improvements

AI Training Capabilities

Blackwell Ultra enables: - GPT-5 class models: Training in weeks vs months - Multimodal models: Efficient vision-language training - Dense models: Up to 10T parameters on single system - Fine-tuning: Hours vs days for large models

Inference Performance

For deployed models: - 4x higher throughput - 60% lower latency - 3x better cost-efficiency - Supports larger batch sizes

Availability and Pricing

Product Availability MSRP
B300 Q2 2026 $40,000
B300X Q3 2026 $35,000
GB300 NVL72 Q4 2026 Contact Sales

Cloud Availability

Major cloud providers offering Blackwell Ultra: - AWS EC2 Ultra instances - Google Cloud A5 instances - Azure ND v5 instances - Oracle Cloud BM.GPU.B300

Competition

The competitive landscape: - AMD MI400: Competitive on price, behind on performance - Intel Gaudi 3: Focus on cost-efficiency - Google TPU v6: Internal use, limited external availability - Custom chips: Meta, Amazon developing in-house

Market Impact

Analyst predictions: - AI chip market to reach $200B by 2027 - NVIDIA share expected to maintain 80%+ of data center GPU market - Shortages expected through 2026 - Supply chain expanding rapidly

Developer Tools

NVIDIA provides: - CUDA 14 with Blackwell optimizations - TensorRT 10 for inference - NeMo framework updates - Containerized development environments

Sustainability

Environmental considerations: - 50% better performance per watt - Liquid cooling options - Carbon offset programs - Energy-efficient data center designs

Blackwell Ultra reinforces NVIDIA's leadership in AI hardware, providing the compute foundation for the next generation of AI models and applications.

Source: Jack AI Hub