TechBriefAI

NVIDIA Details Full-Stack AI Inference and Networking Platform at Hot Chips

Executive Summary

NVIDIA is previewing its upcoming presentations at the Hot Chips conference, where it will showcase a comprehensive platform for AI inference and networking. The announcement details key technologies built on the Blackwell architecture, including the GeForce RTX 5090 GPU, the GB200 NVL72 rack-scale system, and the new DGX Spark desktop supercomputer. Central to this strategy are advanced networking solutions like Spectrum-XGS Ethernet, designed to scale AI workloads from the desktop to multi-data center "AI super-factories."

Key Takeaways

* Full-Stack Showcase: The announcement serves as a preview of NVIDIA's talks at the Hot Chips conference, focusing on its unified platform for accelerating AI inference at every scale.

* Advanced Networking: NVIDIA is highlighting its networking portfolio, including the `ConnectX-8 SuperNIC` for high-speed multi-GPU communication and the new `Spectrum-XGS Ethernet`, a "scale-across" technology to unify distributed data centers. Co-packaged optics (CPO) switches are also featured for enabling more efficient, gigawatt-scale AI factories.

* Blackwell Architecture in Action: The platform is powered by the Blackwell architecture, which is showcased in several products:

* GeForce RTX 5090 GPU: A consumer GPU for next-level graphics and neural rendering.

* GB200 NVL72: An exascale computer in a single rack, featuring 36 GB200 Superchips for massive AI reasoning workloads.

* Desktop Supercomputing: The announcement introduces the `DGX Spark`, a desktop supercomputer powered by the `GB10 Superchip` and designed to bring generative AI development to researchers, data scientists, and students.

* Software and Optimization: The platform introduces `NVFP4`, a low-precision numerical format for more efficient agentic AI inference, particularly for LLMs. NVIDIA also emphasizes its contributions to open-source libraries (e.g., TensorRT-LLM) and its NIM microservices for easy model deployment.

Strategic Importance

This announcement reinforces NVIDIA's strategy of providing an end-to-end, integrated AI platform, aiming to lock in customers across every scale from desktop to interconnected data centers. It solidifies their market dominance by addressing the entire AI workflow, especially the growing demand for large-scale, efficient inference.

Original article