The Business & Technology Network
Helping Business Interpret and Use Technology
«  
  »
S M T W T F S
 
 
 
 
 
 
1
 
2
 
3
 
4
 
5
 
6
 
7
 
8
 
9
 
 
 
 
 
 
 
 
 
 
 
20
 
21
 
22
 
23
 
24
 
25
 
26
 
27
 
28
 
29
 
30
 
31
 
 
 
 
 
 

GTC 2025: The most important AI announcements

DATE POSTED:March 19, 2025
 The most important AI announcements

If you thought AI computing was already moving at breakneck speed, GTC 2025 just hit turbo mode. Nvidia unveiled game-changing advancements in AI hardware, software, and robotics—pushing boundaries in AI reasoning, inference acceleration, and 6G connectivity. With Blackwell Ultra, Dynamo, and GR00T N1, Nvidia isn’t just leading the AI revolution—it’s rewriting the rules. Let’s break down the biggest announcements.

Nvidia RTX Pro Blackwell series: AI and 3D work redefined

Nvidia has officially launched the RTX Pro Blackwell series at GTC 2025, introducing an unprecedented level of power for AI workloads and 3D processing. The flagship RTX Pro 6000 Blackwell for workstations features 96GB of GDDR7 memory, requiring 600 watts of power—an increase from the 575 watts of the RTX 5090. It boasts 24,064 CUDA cores, 512-bit memory bus, and 1792 GB/s memory bandwidth, supporting PCIe Gen 5 and DisplayPort 2.1.

This professional-grade GPU is tailored for game developers, AI professionals, and data-heavy workflows, coming in Max-Q and server variants. Nvidia has rebranded its professional GPU lineup under “RTX Pro,” replacing the previous Quadro series. The series includes RTX Pro 5000, 4000, and 4500 models for desktops and laptops, with up to 24GB of VRAM for laptop variants.

Nvidia RTX Pro Blackwell series pack insane power for AI and 3D work

Key specifications:

  • Fourth-Generation RT Cores for double the photorealistic rendering performance.
  • Fifth-Generation Tensor Cores handling up to 4,000 AI trillion operations per second, with support for FP4 precision and DLSS 4 Multi Frame Generation.
  • GDDR7 Memory scaling up to 96GB for professional workstations.
  • Sixth-Generation NVIDIA NVDEC enhancing video processing.
  • Multi-Instance GPU (MIG) for workload partitioning.

The RTX Pro 6000 will be available in April through PNY and TD Synnex, while OEMs like Dell, HP, and Lenovo will integrate it in May. Cloud providers including AWS, Google Cloud, and Azure will also adopt the GPU later this year.

Nvidia Dynamo: A 30x boost for AI inference

To address performance concerns with emerging AI models like DeepSeek R1, Nvidia unveiled Nvidia Dynamo, a parallel-processing AI inference software. This technology distributes AI workloads across up to 1,000 Nvidia GPUs, significantly increasing query throughput and efficiency.

According to Ian Buck, Nvidia’s head of hyperscale and high-performance computing, Dynamo delivers 30 times more performance using the same GPU count. This efficiency boost translates to 50 times more revenue potential for data centers utilizing Blackwell GPUs over previous Hopper-based architectures.

Additionally, Nvidia optimized DeepSeek R1 by reducing its floating-point precision to FP4, a move that increases efficiency without sacrificing accuracy. The Blackwell Ultra variant, an enhancement of the Blackwell 200, now includes 288GB of HBM3e memory and integrates with Grace CPUs in the NVL72 rack system for 50% better inference performance.

Nvidia can boost DeepSeek R1’s speed 30x, says Jensen Huang

DGX Spark: The world’s smallest AI supercomputer

Nvidia introduced DGX Spark, a $3,000 personal AI supercomputer, marking a new era for desktop AI development. The device, previously codenamed “Digits,” is powered by the GB10 Blackwell Superchip, which delivers 1,000 trillion operations per second (TOPS).

Key specifications:

  • 128GB of unified memory
  • Up to 4TB of NVMe SSD storage
  • Optimized for fine-tuning and AI inference tasks

For those needing higher performance, the DGX Station was also announced, featuring the GB300 Blackwell Ultra desktop superchip, capable of 20 petaflops of AI processing. Nvidia has partnered with Asus, Dell, HP, Boxx, Lambda, and Supermicro to launch variations of these systems.

Nvidia reveals DGX Spark: The world’s smallest AI supercomputer

Blackwell Ultra GB300: Nvidia’s next AI superchip

Expanding its AI hardware dominance, Nvidia unveiled the Blackwell Ultra GB300, shipping in late 2025. Despite not using a new architecture, it enhances the existing Blackwell series with 288GB of HBM3e memory and delivers 1.5x the FP4 inference performance of previous generations.

This chip forms the core of the NVL72 and NVL576 rack systems, dramatically improving AI model efficiency. Nvidia’s next-gen architectures, Vera Rubin (2026) and Rubin Ultra (2027), promise further scaling to 100 petaflops of FP4 performance.

AI at warp speed: Nvidia’s new GB300 superchip arrives this year

GR00T N1: AI-Powered Robotics with Human-Like Cognition

Nvidia’s AI robotics push continues with the GR00T N1 foundation model, designed to mimic human cognition. It features:

  • System 1 for fast-thinking, reflex-based actions
  • System 2 for methodical, reasoning-based decisions
  • Training on 780,000 synthetic trajectories for enhanced learning

The GR00T N1 model powers 1X Technologies’ humanoid robots, performing real-world tasks like household chores. Nvidia also introduced Newton, a physics engine co-developed with Google DeepMind and Disney Research, enabling precise robotic interactions.

GR00T N1: Nvidia’s next leap in AI-powered robotics

Llama Nemotron: Nvidia’s open source reasoning model

Nvidia responded to the AI reasoning boom with Llama Nemotron, an open-source alternative to models like DeepSeek R1. The model family includes:

  • Nemotron Nano (for edge deployments)
  • Nemotron Super (for high-throughput inference)
  • Nemotron Ultra (for multi-GPU reasoning)

A notable feature is its adaptive reasoning toggle, allowing AI systems to switch between full reasoning and direct response modes, optimizing performance and cost efficiency. Additionally, Nvidia introduced Agent AI-Q, an open-source framework for integrating AI reasoning models into enterprise workflows.

Llama Nemotron: Nvidia’s answer to the AI reasoning boom

AI-RAN and the future of 6G wireless

Nvidia is accelerating AI-native 6G networks with its Aerial Research portfolio, which includes:

  • Aerial Omniverse Digital Twin (AODT) for real-world simulation
  • Aerial Commercial Test Bed (ARC-OTA) for over-the-air AI model deployment
  • Sionna 1.0, an open-source library for AI-driven radio communication modeling

Telecom giants like SoftBank and Fujitsu reported 50% throughput increases in challenging radio conditions using AI-based uplink channel interpolation. Nvidia’s AI-RAN technology is set to revolutionize the way 6G networks are built and optimized.

Nvidia’s AI-RAN tech is building the future of 6G wireless

GTC 2025 ushers in an AI-powered future

Nvidia’s GTC 2025 showcased a sweeping transformation in AI hardware, from Blackwell Ultra GPUs to Dynamo software and AI-driven wireless networks. The launch of DGX Spark, Llama Nemotron models, and AI-powered robotics reinforces Nvidia’s dominant position in AI computing. As these technologies roll out throughout 2025 and beyond, they will redefine AI inference, reasoning, and robotics across industries.

Featured image credit: Nvidia