Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

NVIDIA Infrastructure

Research Note: NVIDIA SC24 Announcements

At the recent Supercomputing 2024 (SC24) conference in Atlanta, NVIDIA announced new hardware and software capabilities to enhance AI and HPC capabilities. This includes the new GB200 NVL4 Superchip, the general available of its H200 NVL PCIe, and several new software capabilities.

Hardware Updates

NVIDIA made hardware announcements that reinforce its leadership in AI and HPC, including its new GB200 NVL4 Superchip and the general availability of its Hopper-based NVL PCIe.

GB200 NVL4 Superchip:

The new GB200 NVL4 Superchip integrates two Grace CPUs and four Blackwell B200 GPUs on a single board, interconnected via NVIDIA’s NVLink technology. The architecture provides 1.3TB of coherent HBM3 memory shared across the GPUs.

NVIDIA says that the GB200 NVL4 offers substantial improvements over its predecessor, the GH200 NVL4, delivering 2.2 times faster simulation performance and 1.8 times faster training and inference capabilities.

The new part is expected to be available to customers in the second half of 2025.

H200 NVL PCIe Availability:

NVIDIA also announced the general availability of the Hopper generation H200 NVL PCIe GPU. This GPU is built for data centers with lower power requirements and air-cooled enterprise rack designs.

Compared to its predecessor, the H100 NVL, the H200 NVL provides a 1.5x increase in memory and a 1.2x boost in bandwidth. These improvements enable up to 1.7x faster inference performance for large language models and a 1.3x performance increase for HPC applications.

The H200 NVL comes with a five-year subscription to NVIDIA AI Enterprise, a cloud-native software platform for developing and deploying production AI.

Software Updates

At the SC24 event, NVIDIA unveiled several software and platform developments to advance AI and HPC across various industries.

cuPyNumeric Library

NVIDIA introduced cuPyNumeric, a GPU-accelerated implementation of NumPy, designed to enhance applications in data science, machine learning, and numerical computing. This addition expands NVIDIA’s suite of over 400 CUDA-X libraries.

Omniverse Blueprint for Real-Time Digital Twins

The company announced the NVIDIA Omniverse Blueprint, a reference workflow that enables developers to create interactive digital twins for sectors such as aerospace, automotive, energy, and manufacturing. This blueprint integrates NVIDIA’s acceleration libraries, physics-AI frameworks, and interactive physically based rendering, achieving up to 1,200x faster simulations and real-time visualization.

Earth-2 NIM Microservices

For climate forecasting, NVIDIA introduced two new NVIDIA NIM microservices for its Earth-2 digital twin platform. The CorrDiff NIM and FourCastNet NIM microservices can accelerate climate change modeling and simulation results by up to 500x, enhancing the ability to simulate and visualize weather and climate conditions.

Analysis

NVIDIA has always used software as an enabler to further its hardware solutions.   The company’s dual focus on high-performance hardware and comprehensive software continued at SC24, part of its calculated strategy to maintain its competitive edge.

On the hardware side, the H200 NVL PCIe’s availability broadens NVIDIA’s reach into enterprise data centers with a modular, more power-efficient solution. While it doesn’t match the raw power of the GB200, its flexibility and cost efficiency will appeal to a broader range of customers.

NVIDIA’s primary challenge remains power consumption, particularly with the GB200 NVL4 requiring 5,400 watts. This raises valid questions about sustainability and deployment feasibility, especially for enterprises with constrained cooling infrastructure.

NVIDIA continues to be the “go-to” provider of comprehensive AI and HPC solutions by integrating cutting-edge hardware with advanced software platforms. This approach strengthens its ecosystem, fosters customer loyalty, and expands its influence across sectors.

Competitive Differentiation & Advice to IT Buyers

These sections are only available to clients. Please reach out to info@nand-research.com to learn more.

Disclosure: The author is an industry analyst, and NAND Research an industry analyst firm, that engages in, or has engaged in, research, analysis, and advisory services with many technology companies, which may include those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.