STOCK TITAN

HPE Unveils Next-Generation AI Factory and Supercomputing Advancements with NVIDIA

Rhea-AI Impact
(Moderate)
Rhea-AI Sentiment
(Very Positive)
Tags
AI

Key Terms

infiniBand technical
Infiniband is a high-speed data transport technology used inside data centers to move large amounts of information quickly and with very little delay, often used for servers, storage and computing clusters. Investors should care because it acts like a multilane expressway for data—companies that build, use or support such fast networks can gain competitive advantages in cloud services, high-performance computing and AI workloads, which can affect costs and revenue potential.
liquid cooling technical
Liquid cooling is a method that uses a flowing liquid—like water or a special coolant—to carry heat away from electronic components, similar to how a car radiator moves heat away from an engine. For investors, it matters because it can lower energy and maintenance costs, enable higher-performance computing, reduce the footprint of data centers, and support sustainability targets, all of which can affect a company’s operating margins and capital spending needs.
multi-tenancy technical
Multiple customers or user groups share the same software application and underlying infrastructure while keeping their data and settings separate, like different businesses renting rooms in the same office building but using shared utilities. For investors, multi-tenancy matters because it lets a software provider add customers faster and at lower cost, improving margins and growth potential, but also concentrates operational and security risks that can affect revenue and reputation.
virtual machines technical
A virtual machine is a software-created computer that runs inside a physical server, allowing one piece of hardware to behave like many separate computers. For investors, virtual machines matter because they enable companies to scale computing capacity, reduce costs, and deploy new services quickly—similar to renting multiple apartments inside a single building instead of buying separate houses, which can improve efficiency and influence profitability and capital needs.
kubernetes technical
Kubernetes is an open-source system that automates running and managing many pieces of software across groups of computers, like a conductor coordinating musicians so each piece plays at the right time and place. For investors, it matters because companies that use it can deploy updates faster, scale services up or down automatically, and cut infrastructure costs — factors that influence growth, reliability and operating margins.
multi-instance gpu technical
A multi-instance GPU lets a single physical graphics processor be split into several independent, smaller GPUs so different tasks or customers can use the same hardware at once. For investors, it means data centers and cloud providers can squeeze more revenue and efficiency from expensive GPUs—like turning one large office into multiple rentable cubicles—reducing costs and improving flexibility for AI, graphics, and compute workloads.
open compute project technical
An industry collaboration that creates and shares open designs for data-center hardware—like servers, storage and networking—so companies can build or buy compatible, efficient equipment without proprietary lock-in. For investors, it matters because those shared designs can lower capital and operating costs, speed innovation, and increase competition among suppliers, all of which can influence profit margins and capital spending needs for technology-focused businesses.
data sovereignty regulatory
Data sovereignty is the principle that digital information is subject to the laws and control of the country or entity where it is stored or processed. For investors, it matters because where data lives affects a company's legal obligations, costs, ability to sell services across borders, and exposure to government access or restrictions — like owning a house that must follow the rules of the town it sits in.

New compute and networking offerings pair with full-stack software and services to accelerate AI adoption across service providers, sovereign entities, research labs, and large enterprises

In this article

  • HPE is first to offer a NVIDIA Vera CPU compute blade and adds NVIDIA Quantum-X800 InfiniBand networking to the leadership-class HPE Cray supercomputers.
  • HPE AI Factory portfolio now features NVIDIA Vera Rubin and NVIDIA Blackwell platforms, bolstering performance for service providers and sovereign entities.
  • New offerings include NVIDIA Cloud Partner endorsement, multi-tenancy through NVIDIA Multi-Instance GPU, support for NVIDIA Mission Control, and Red Hat with NVIDIA software.

SAN JOSE, Calif.--(BUSINESS WIRE)-- NVIDIA GTC 2026 - HPE (NYSE: HPE) today announced significant innovations to the NVIDIA AI Computing by HPE portfolio focused on large-scale AI factories and supercomputers that enable customers to scale, deploy efficiently, and gain faster time-to-insight. The full-stack AI solutions with NVIDIA include tightly integrated compute, GPUs, networking, liquid cooling, software, and services designed for at-scale and sovereign environments. AI-forward organizations and leading research institutions, including Argonne National Laboratory, HLRS, Hudson River Trading (HRT), and the Korea Institute of Science and Technology Information (KISTI), have chosen HPE AI infrastructure and AI factories with NVIDIA to advance innovation.

HPE brings NVIDIA AI solutions to its industry-leading supercomputing platform

Research laboratories, sovereign entities and large enterprises are rapidly adopting AI to enhance traditional high performance computing (HPC) workloads. For organizations seeking to significantly expedite scientific discovery, HPE is making the following NVIDIA products available on its second-generation exascale-class supercomputing platform designed to unify AI and HPC – the HPE Cray Supercomputing GX5000.

  • HPE introduces one of the first NVIDIA Vera CPU compute blades – HPE adds a new liquid-cooled NVIDIA compute blade option to its lineup of powerful, next-generation supercomputing solutions. Each HPE Cray Supercomputing GX240 Compute blade features up to 16 NVIDIA Vera CPUs and is designed to support the most demanding AI compute workloads. The GX240 offers industry-leading density1 on the NVIDIA Vera platform, scaling up to 40 blades with 640 NVIDIA Vera CPUs, and 56,320 NVIDIA Olympus Arm-compatible cores per rack.
  • More choices for high-performance networking – Organizations can customize their supercomputers with a choice of networking options optimized for large-scale systems that includes NVIDIA Quantum-X800 InfiniBand. Now available with HPE Cray Supercomputing GX5000, NVIDIA Quantum-X800 InfiniBand switches provide 144 ports of 800 Gb/s connectivity per port with power efficiency features, including low-power link state and power profiling.

“Having built the three most powerful, exascale supercomputers in the world, HPE is at the forefront of innovation that brings together cutting-edge AI workloads with traditional HPC to accelerate scientific breakthroughs,” said Trish Damkroger, senior vice president and general manager, HPC & AI Infrastructure Solutions at HPE. “Our continued collaboration with NVIDIA helps customers tap into the high-performance density they need to push the boundaries in the fields of medicine, life sciences, engineering, manufacturing, and more.”

Enhancing HPE AI Factory at-scale and sovereign, co-engineered with NVIDIA

In addition to the enhancements of its industry-leading supercomputing platform2, HPE is bolstering the HPE AI Factory portfolio for service providers, sovereign entities, and large enterprises with the NVIDIA Vera Rubin platform and NVIDIA Blackwell architecture.

  • At-scale AI deployments for neo-clouds – HPE introduces the next-generation NVIDIA Vera Rubin NVL72 by HPE rack-scale system, a flagship AI system engineered for frontier‑scale models in excess of 1 trillion parameters. This system, designed for neo-clouds, delivers high efficiency at-scale with 36 NVIDIA Vera CPUs, 72 NVIDIA Rubin GPUs, sixth generation NVIDIA NVLink scale-up networking, NVIDIA ConnectX‑9 SuperNIC, and NVIDIA BlueField‑4 DPUs along with HPE’s liquid cooling integration, services, and expertise for data center design3 that eases large-scale AI rollouts.
  • High-density GPU server for AI model training and inference – The HPE Compute XD700 is a new Open Compute Project (OCP)‑inspired4 AI server built on NVIDIA HGX Rubin NVL8. The system is designed to deliver higher GPU density per rack and reduce space, power, and cooling costs while increasing AI training and inference throughput. Each rack of XD700 servers supports up to 128 Rubin GPUs, providing double the GPU density compared to the previous generation5.
  • More NVIDIA Blackwell access – The NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs are newly available across all of HPE’s AI factories.

These solutions are enhanced by a number of HPE and NVIDIA software and services updates that lead to faster AI deployments for customers rolling out large-scale AI projects.

  • HPE AI Factory portfolio is NVIDIA Cloud Partner program-endorsed – Through the ongoing engineering collaboration with NVIDIA, HPE’s AI factories are ready for NVIDIA Cloud Provider certification, enabling cloud service providers to streamline the validation process.
  • Expanding multi-tenancy options for at-scale AI deployments – HPE enhances the HPE AI Factory portfolio by supporting multi-tenancy models for virtual machines (VMs) with GPU passthrough and secure Kubernetes namespaces through NVIDIA Multi-Instance GPU (MIG) enabled by SUSE Virtualization and SUSE Rancher Prime Suite. This allows service providers to choose between hard and soft tenancy deployment models to fit customer needs.
  • Red Hat integration – The HPE AI Factory portfolio currently supports Red Hat Enterprise Linux and OpenShift, part of Red Hat AI Enterprise, which integrates with NVIDIA AI Enterprise solutions for customers seeking an enterprise Linux operating system.
  • HPE AI Factory with Mission Control – HPE AI Factory at scale and HPE AI Factory sovereign will offer NVIDIA Mission Control software, streamlining every aspect of the AI factory from workload orchestration with NVIDIA Run:ai to NVIDIA Dynamo that addresses monitoring and autonomous recovery – while empowering platform teams to operate efficiently and scale confidently.

Each of these solutions are built leveraging HPE’s services and expertise in datacenter design and liquid cooling6 gained through decades of experience building the largest and most energy-efficient supercomputers in the world7.

“To realize the potential of AI, enterprises and nations require infrastructure that can handle massive-scale model training and HPC workloads,” said Chris Marriott, vice president, Enterprise Platforms at NVIDIA. “Together, HPE and NVIDIA have developed full-stack AI infrastructure that unite accelerated computing, advanced networking and liquid cooling for faster time-to-insight in at-scale and sovereign environments.”

Availability

  • HPE Cray Supercomputing GX240 Compute blade with up to 16 NVIDIA Vera CPUs will be available in 2027.
  • NVIDIA Quantum-X800 InfiniBand networking is available on HPE Cray Supercomputing GX5000 in 2027.
  • The NVIDIA Vera Rubin NVL72 by HPE rack-scale system will be available in December 2026.
  • The HPE Compute XD700 will be available in early 2027.
  • NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs are available in the HPE AI Factory portfolio today.
  • HPE AI Factory portfolio with multi-scale tenancy and GPU passthrough will be available in Spring 2026.
  • Red Hat Enterprise Linux and Red Hat OpenShift integration with NVIDIA is available in HPE AI Factory portfolio today.
  • HPE AI Factory at scale and sovereign support of NVIDIA Mission Control software is planned for 2026.

Read more about HPE and NVIDIA’s sovereign AI customer stories here

Additional Resources

About HPE

HPE (NYSE: HPE) is a leader in essential enterprise technology, bringing together the power of AI, cloud, and networking to help organizations achieve more. As pioneers of possibility, our innovation and expertise advance the way people live and work. We empower our customers across industries to optimize operational performance, transform data into foresight, and maximize their impact. Unlock your boldest ambitions with HPE. Discover more at www.hpe.com.

1 HPE internal market research, March 2026

2 HPE has built and delivered the world’s three fastest exascale supercomputers as of November 2025, as validated by the TOP500

3 Top 80 Data Center Engineering Firms for 2025, BDE, October 2025

4 The Open Compute Project (OCP) defines OCP-inspired products as those that comply with approved OCP specifications or contributions

5 Compared to the HPE ProLiant Compute XD685

6 HPE has more than 300 patents and five decades of experience in direct liquid cooling

7 HPE has built and delivered 6 of the 10 fastest supercomputers in the world and 10 of the 20 most energy-efficient supercomputers in the world as validated by November 2025 TOP500 and Green500 lists

Media Contact:

Cristina Thai

cristina.thai@hpe.com

Source: Hewlett Packard Enterprise

Hewlett Packard Enterprise Co

NYSE:HPE

View HPE Stock Overview

HPE Rankings

HPE Latest News

HPE Latest SEC Filings

HPE Stock Data

28.99B
1.32B
Communication Equipment
Computer & Office Equipment
Link
United States
SPRING