The AI race explodes as HPE deploys AMD’s Helios racks, crushing limits with Venice CPUs and insane GPU density
HPE and AMD bring Helios to commercial AI systems
- HPE will ship 72-GPU racks with next-generation AMD Instinct accelerators globally
- Venice CPUs paired with GPUs target exascale-level AI performance per rack
- Helios relies on liquid cooling and double-wide chassis for thermal management
HPE has announced plans to integrate AMD’s Helios rack-scale AI architecture into its product lineup starting in 2026.
The collaboration gives Helios its first major OEM partner and positions HPE to ship full 72-GPU AI racks built around AMD’s next-generation Instinct MI455X accelerators.
These racks will pair with EPYC Venice CPUs and use an Ethernet-based scale-up fabric developed with Broadcom.
Rack layout and performance targets
The move creates a clear commercial route for Helios and puts the architecture in direct competition with Nvidia’s rack-scale platforms already in service.
The Helios reference design relies on Meta’s Open Rack Wide standard.
It uses a double-wide liquid-cooled chassis to house the MI450-series GPUs, Venice CPUs, and Pensando networking hardware.
AMD targets up to 2.9 exaFLOPS of FP4 compute per rack with the MI455X generation, along with 31TB of HBM4 memory.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
The system presents every GPU as part of a single pod, which allows workloads to span all accelerators without local bottlenecks.
A purpose-built HPE Juniper switch supporting Ultra Accelerator Link over Ethernet forms the high-bandwidth GPU interconnect.
It offers an alternative to Nvidia’s NVLink-centric approach.
The High-Performance Computing Center Stuttgart has selected HPE’s Cray GX5000 platform for its next flagship system, named Herder.
Herder will use MI430X GPUs and Venice CPUs across direct liquid-cooled blades and will replace the current Hunter system in 2027.
HPE stated that the GX5000 racks’ waste heat will warm campus buildings, which shows environmental considerations in addition to performance goals.
AMD and HPE plan to make Helios-based systems globally available next year, expanding access to rack-scale AI hardware for research institutions and enterprises.
Helios uses an Ethernet fabric to connect GPUs and CPUs, which contrasts with Nvidia’s NVLink approach.
The use of Ultra Accelerator Link over Ethernet and Ultra Ethernet Consortium-aligned hardware supports scale-out designs within an open standards framework.
Although this approach allows theoretically comparable GPU counts to other high-end AI racks, performance under sustained multi-node workloads remains untested.
However, reliance on a single Ethernet layer could introduce latency or bandwidth constraints in real applications.
That said, these specifications do not predict real-world performance, which will depend on effective cooling, network traffic handling, and software optimization.
Via Tom's Hardware
Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!
And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.

Efosa has been writing about technology for over 7 years, initially driven by curiosity but now fueled by a strong passion for the field. He holds both a Master's and a PhD in sciences, which provided him with a solid foundation in analytical thinking.
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.