In its bid to reduce the barrier of entry for high performance computing, Nvidia (opens in new tab) has announced two new GPUs (opens in new tab) that will happily plug into both existing enterprise servers (opens in new tab) and workstations (opens in new tab).
The A10 and A30 GPUs are both based on the Ampere architecture and will slot in one or two full-length PCIe ports, respectively.
Speaking to The Register, Manuvir Das, Nvidia’s head of enterprise computing, reasoned that the cards are the manifestation of Nvidia’s attempts to give users the means to tackle Artificial Intelligence (AI) workloads on typical datacenter hardware and racks.
- These are the best cloud computing (opens in new tab) services
- We’ve also rounded up the best cloud analytics platforms (opens in new tab)
- Here’s our selection of the best cloud hosting (opens in new tab) providers
Das further assures that Nvidia’s Kubernetes-based EGX platform for machine learning at the edge of the network will work perfectly well with the new GPUs.
Reportedly, Nvidia has been able to tie up with many of the big server-makers including Dell (opens in new tab), Lenovo (opens in new tab), H3C, Inspur, QCT and Supermicro to package the A10 and A30 inside their enterprise and datacenter offerings.
The new GPUs cost about $2000 and $3000 respectively. The A10 is rated at 150W TDP and thanks to 24GB GDDR6 memory can produce 31.2 single-precision teraFLOPS. On the other hand, the A30 is rated at 165W and with the help of 24GB high bandwidth memory v2 delivers 10.3 single-precision teraFLOPS.
While the A10 will go on sale later this month, the A30 is expected to launch sometime later this year.
- We've also featured the best cloud management software (opens in new tab)
Via: The Register (opens in new tab)