Emphasising cost reduction and sustainability, Nvidia founder and CEO Jensen Huang has detailed new semiconductors, software and systems to power data centres, factories, consumer devices, robots and more, with the aim of driving a new industrial revolution.
“Generative AI is reshaping industries and opening new opportunities for innovation and growth,” Huang says. “Today, we’re at the cusp of a major shift in computing. The intersection of AI and accelerated computing is set to redefine the future.
“The future of computing is accelerated,” Huang said. “With our innovations in AI and accelerated computing, we’re pushing the boundaries of what’s possible and driving the next wave of technological advancement.”
Huang revealed a roadmap for new semiconductors.
The Rubin platform will succeed the upcoming Blackwell platform, featuring new GPUs, a new Arm-based CPU – Vera – and advanced networking with NVLink 6, CX9 SuperNIC and the X1600 converged InfiniBand/Ethernet switch.
Nvidia is driving down the cost of turning data into intelligence, Huang explained.
“Accelerated computing is sustainable computing,” he emphasised, outlining how the combination of GPUs and CPUs can deliver up to a 100x speedup while only increasing power consumption by a factor of three, achieving 25x more performance per Watt over CPUs alone.
“The more you buy, the more you save.”
The Nvidia MGX modular reference design platform now supports Blackwell, including the GB200 NVL2 platform, designed for optimal performance in large language model inference, retrieval-augmented generation and data processing.
In networking, Huang has unveiled plans for the annual release of Spectrum-X products to cater to the growing demand for high-performance Ethernet networking for AI.
Nvidia Spectrum-X, the first Ethernet fabric built for AI, enhances network performance by 1,6x more than traditional Ethernet fabrics. It accelerates the processing, analysis and execution of AI workloads and, in turn, the development and deployment of AI solutions.
With Nvidia NIM, the world’s 28-million developers can now easily create generative AI applications. NIM – inference microservices that provide models as optimized containers – can be deployed on clouds, data centres or workstations.
NIM also enables enterprises to maximise their infrastructure investments. For example, running Meta Llama 3-8B in a NIM produces up to 3x more generative AI tokens on accelerated infrastructure than without NIM.
Nvidia’s RTX AI PCs, powered by RTX technologies, will enhance consumer experiences with over 200 RTX AI laptops and more than 500 AI-powered apps and games.
Project G-Assist, an RTX-powered AI assistant technology demo, was also announced, showcasing context-aware assistance for PC games and apps.
Microsoft and Nvidia are collaborating to help developers bring new generative AI capabilities to their Windows native and web apps with easy API access to RTX-accelerated SLMs that enable RAG capabilities that run on-device as part of Windows Copilot Runtime.
According to Huang, Nvidia is spearheading the $50-trillion industrial digitisation shift, with sectors embracing autonomous operations and digital twins – virtual models that enhance efficiency and cut costs. Through its Developer Program, Nvidia offers access to NIM.