At GTC 2025 NVIDIA unveiled the Blackwell Ultra GPU and Rubin chips and servers

NVIDIA’s GTC 2025 event introduced cutting-edge AI hardware, software, and partnerships, emphasizing scalable infrastructure for AI reasoning, industrial automation, and enterprise adoption, while expanding its AI ecosystem with open-source reasoning models and agentic AI tools.

At GTC 2025 NVIDIA unveiled the Blackwell Ultra GPU and Rubin chips and servers - Credit - Nvidia, Midjourney, The AI Track
At GTC 2025 NVIDIA unveiled the Blackwell Ultra GPU and Rubin chips and servers - Credit - Nvidia, Midjourney, The AI Track

NVIDIA announcements at GTC 2025 – Key Points

  1. GPU Roadmap Expansion:
    • Blackwell Ultra NVL72 GPU: Launched for late 2025, delivers 20 petaflops and 1.5x performance over the original Blackwell, targeting AI reasoning workloads. Offers 50x greater data center revenue opportunity vs. Hopper chips.
    • GB300 Superchip: Combines two Blackwell Ultra GPUs with NVIDIA’s Grace CPU, powering hyperscalers (Amazon, Google, Microsoft, Meta) and research labs.
    • Performance Metrics:
      • Processes 1,000 tokens/second with DeepSeek’s R1 model (vs. 100 tokens/second on Hopper).
      • Reduces query response time to 10 seconds (vs. 1.5 minutes on Hopper).
    • Vera Rubin GPU: Confirmed for 2026 with 50 petaflops, targeting generative AI models.
    • Rubin Ultra GPU: Planned for 2027 at 100 petaflops, addressing exponential AI demands.
    • Feynman GPUs (2028): Extend NVIDIA’s roadmap for AI hardware dominance.
  2. DGX Systems & SuperPods:
    • DGX Spark: Immediate availability with GB10 Superchip (1,000 trillion ops/sec), designed for edge AI prototyping.
    • DGX Station: Features GB300 Superchip (784GB memory), launching via Asus, Dell, HP, Lenovo in late 2025.
    • DGX SuperPod: Combines 288 Grace CPUs + 576 Blackwell Ultra GPUs + 300TB memory, forming enterprise-scale AI powerhouses.
  3. Nvidia Dynamo: Open-source OS for AI factories, optimizing inference efficiency for reasoning models to reduce deployment costs.
  4. Robotics & Simulation:
    • Project Groot N1: Open-source AI foundation model for humanoid robots, featuring a dual-system architecture (“fast” and “slow” thinking) inspired by human cognition.
      • Slow System: Perceives environments, reasons about tasks, and plans actions.
      • Fast System: Executes multi-step physical manipulations.
      • Evolved from Project Groot (industrial-focused predecessor) to support diverse environments.
      • Trained on synthetic + real-world data, with simulation frameworks and synthetic data blueprints released alongside.
    • Omniverse Blueprint: Enhanced simulation-first platform for digital twin development in industrial automation.
  5. Industry Partnerships:
    • Hyperscalers: Blackwell Ultra adopted by Amazon, Google, Microsoft, Meta for AI reasoning systems.
    • General Motors (GM): Expanded collaboration to integrate AI into self-driving cars, manufacturing, and design processes.
    • Disney, Google & NVIDIA: Joint development of Newton, a physics engine for real-time robotic movement simulations, to power Disney’s next-gen entertainment robots by 2026.
    • Enterprise Collaborators: Accenture, SAP, Microsoft, Deloitte, ServiceNow, and others leveraging NVIDIA’s new reasoning models for industry-specific AI agents.
  6. Software Ecosystem:
    • Llama Nemotron Reasoning Models: Open-source family (Nano-8B, Super-49B, Ultra-249B) optimized for agentic AI:
      • Performance: 20% higher accuracy vs. base Llama models; 5x faster inference than competitors.
      • Toggle Feature: Switches between intensive reasoning and direct responses.
    • AI-Q Blueprint: Framework (April 2025 release) to connect AI agents with enterprise systems/data.
    • NIM Microservices: Optimized inference tools for agentic AI, integrated with Azure AI Foundry, SAP Business AI, and ServiceNow workflows.
    • CUDA & AI Enterprise Suite: Updates for generative AI and reasoning workflows.
    • Omniverse Blueprint: Enhanced tools for robotics and automation.
  7. Financial & Market Impact:
    • Blackwell Revenue: Contributed $11 billion to NVIDIA’s Q1 2025 total revenue of $39.3 billion.
    • Fastest Production Ramp: Blackwell is NVIDIA’s most rapidly scaled chip, now in full production.
    • Stock Reaction: Shares fell 3% post-announcements, reflecting investor skepticism over long-term, “unsurprising” updates.

Why This Matters:

NVIDIA’s advancements at GTC 2025 solidify its role as the backbone of enterprise and industrial AI. The Llama Nemotron models and AI-Q Blueprint democratize agentic AI development, enabling businesses to deploy autonomous systems for complex problem-solving. Despite technological strides—like the Blackwell Ultra’s 50x performance leap and Groot N1’s cognitive architecture—the lukewarm market reaction underscores challenges in sustaining investor enthusiasm amid high expectations. Partnerships with Disney, GM, SAP, and Microsoft highlight real-world applications, but skepticism persists about scalability and ROI in humanoid robotics and enterprise AI.

Explore the vital role of AI chips in driving the AI revolution, from semiconductors to processors: key players, market dynamics, and future implications.

Read a comprehensive monthly roundup of the latest AI news!

The AI Track News: In-Depth And Concise

Scroll to Top