Alongside an Nvidia-powered version, customers can soon snap up a new version of the Dell PowerEdge XE9680 fitted with eight AMD Instinct MI300X accelerators. This helps businesses train and run their own in-house large language models (LLMs) – offering 1.5GB of high-bandwidth memory (HBM3) and over 21 petaFLOPS of performance.
Customers can also scale systems they deploy with the global memory interconnect (xGMI) standard, and connect AMD's GPUs over an Ethernet-based AI fabric with a Dell PowerSwitch Z9664F-ON. It follows Dell's launch of the unit fitted with Nvidia H100 GPUs earlier this year.
Dell adopts an open approach to AI infrastructure
This comes alongside a new standard called Dell Validated Design for Generative AI with AMD. It's a mouthful, but is essentially a framework for organizations wishing to run their own hardware and networking architecture to run LLMs.
Dell customers could use the solution to build generative AI platforms with various frameworks that guide the integration of the technology as well as physical installation and optimization for performance.
It also uses AMD ROCm powered AI frameworks, which is an open source package of drivers, dev toolkits and APIs compatible with AMD Instinct accelerators. They include popular frameworks like PyTorch, TensorFlow and OpenAI Triton – which all have native support on the PowerEdge XE9680 that's fitted with the AMD accelerators.
Dell's continued push for standards-based networking, as a member of the Ultra Ethernet Consuortium (UEC), means it's adopted a more open approach to that adopted by Nvidia.
Although an AI infrastructure industry leader, Nvidia differs from AMD in that the latter advocates for an open Ethernet for AI – in which switches from different vendors can interoperate together on the same system. Likewise, Dell wants businesses to take an open approach across the compute, fabric as well as storage components needed to power generative AI models in-house.
The new hardware and services that comprise Dell's latest AI push are set to be available during the first half of next year, the company said.
More from TechRadar Pro
Are you a pro? Subscribe to our newsletter
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
Keumars Afifi-Sabet is the Technology Editor for Live Science. He has written for a variety of publications including ITPro, The Week Digital and ComputerActive. He has worked as a technology journalist for more than five years, having previously held the role of features editor with ITPro. In his previous role, he oversaw the commissioning and publishing of long form in areas including AI, cyber security, cloud computing and digital transformation.
IBM has an AI SSD that can detect and eradicate ransomware in 60 seconds — but you'd kid yourself if you think you can buy it and plug it in your PC
Researcher discovers novel way to double computer speeds for free — but there is an obvious catch and it may only work with Nvidia GPU and Arm CPU for now