Dell cosies up with AMD to expand its generative AI portfolio to offer more choice to customers — it’s also embracing standards-based networking, a thinly veiled dig at Nvidia

Dell’s PowerEdge XE9680 system features eight AMD Instinct MI300X AI accelerators each

When you purchase through links on our site, we may earn an affiliate commission.Here’s how it works.

Dellhas added anAMD-powered server to its ranks of high-performance computing portfolio for AI workloads.

Alongside anNvidia-powered version, customers can soon snap up a new version of the Dell PowerEdge XE9680 fitted with eight AMD Instinct MI300X accelerators. This helps businesses train and run their own in-house large language models (LLMs) – offering 1.5GB of high-bandwidth memory (HBM3) and over 21 petaFLOPS of performance.

Customers can also scale systems they deploy with the global memory interconnect (xGMI) standard, and connect AMD’s GPUs over an Ethernet-based AI fabric with a Dell PowerSwitch Z9664F-ON. It followsDell’s launch of the unit fitted with Nvidia H100 GPUsearlier this year.

Dell adopts an open approach to AI infrastructure

Dell adopts an open approach to AI infrastructure

This comes alongside a new standard called Dell Validated Design for Generative AI with AMD. It’s a mouthful, but is essentially a framework for organizations wishing to run their own hardware and networking architecture to run LLMs.

Dell customers could use the solution to build generative AI platforms with various frameworks that guide the integration of the technology as well as physical installation and optimization for performance.

It also uses AMD ROCm powered AI frameworks, which is an open source package of drivers, dev toolkits and APIs compatible with AMD Instinct accelerators. They include popular frameworks like PyTorch, TensorFlow andOpenAITriton – which all have native support on the PowerEdge XE9680 that’s fitted with the AMD accelerators.

Dell’s continued push for standards-based networking, as a member of the Ultra Ethernet Consuortium (UEC), means it’s adopted a more open approach to that adopted by Nvidia.

Are you a pro? Subscribe to our newsletter

Are you a pro? Subscribe to our newsletter

Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!

Although an AI infrastructure industry leader,Nvidia differs from AMDin that the latter advocates for an open Ethernet for AI – in which switches from different vendors can interoperate together on the same system. Likewise, Dell wants businesses to take an open approach across the compute, fabric as well as storage components needed to power generative AI models in-house.

The new hardware and services that comprise Dell’s latest AI push are set to be available during the first half of next year, the company said.

More from TechRadar Pro

Keumars Afifi-Sabet is the Technology Editor for Live Science. He has written for a variety of publications including ITPro, The Week Digital and ComputerActive. He has worked as a technology journalist for more than five years, having previously held the role of features editor with ITPro. In his previous role, he oversaw the commissioning and publishing of long form in areas including AI, cyber security, cloud computing and digital transformation.

AMD just outsold Intel in the data center space for the first time ever

The UK government wants to help businesses make trustworthy AI products

Apple MacBook Pro 14-inch M4 (2024) review: one of the best Pro laptops around just got better