New AI servers can boost performance into supercomputing territory with a smaller footprint. Credit: dny59 / kentoh / Getty Images Graphcore, the British semiconductor company that develops accelerators for AI and machine learning, has greatly increased the performance of its massively parallel Intelligence Processing Unit (IPU) servers. Graphcore sells its AI-oriented chips in rack-mounted designs called IPU-PODs. Up until now, the maximum per cabinet has been 64 units (in the IPU-POD64). The newest racks are twice and four times as large: the IPU-POD128 and IPU-POD256. With 32 petaFLOPS of AI compute in the IPU-POD 128 and 64 petaFLOPS in the IPU-POD 256, Graphcore says it can now extend its reach into AI supercomputer territory, and with a much smaller footprint than the typical supercomputer, which could fill a basketball court. The IPU-POD systems disaggregate the AI compute from the servers, which means different types of AI workloads requiring different levels of performance can all be run on the same POD. For example, a POD can be allocated for faster training of large Transformer-based language models across an entire system, or the system can be divided into smaller, flexible vPODs to give more developers IPU access. This is done through Graphcore’s Poplar software stack, which includes an SDK, the Graphcore Communication Library (GCL) for managing communication and synchronization between IPUs, and PopRun and PopDist, which allow developers to run their applications across multiple IPU-POD systems. For intra-rack IPU communication, earlier PODs used 64Gb/s IPU-Links. Now, the IPU-POD128 and IPU-POD256 use new Gateway Links, a horizontal, rack-to-rack connection that extends IPU-Links using tunneling over regular 100Gb Ethernet. Both systems have been developed for cloud hyperscalers, national scientific computing labs and enterprise companies with large AI teams in markets like financial services or pharmaceutical. Graphcore’s initial customers include the U.S. Department of Energy’s Lawrence Berkeley National Laboratory and J.P. Morgan. IPU-POD16, IPU-POD64, IPU-POD128 and IPU-POD256 are shipping to customers today from French IT giant ATOS and other systems integrator partners around the world and are available to buy in the cloud from Cirrascale. Related content news AMD holds steady against Intel in Q1 x86 processor shipments finally realigned with typical seasonal trends for client and server processors, according to Mercury Research. By Andy Patrizio May 22, 2024 4 mins CPUs and Processors Data Center news Broadcom launches 400G Ethernet adapters The highly scalable, low-power 400G PCIe Gen 5.0 Ethernet adapters are designed for AI in the data center. By Andy Patrizio May 21, 2024 3 mins CPUs and Processors Networking news HPE updates block storage services The company adds new storage controller support as well as AWS. By Andy Patrizio May 20, 2024 3 mins Enterprise Storage Data Center news ZutaCore launches liquid cooling for advanced Nvidia chips The HyperCool direct-to-chip system from ZutaCore is designed to cool up to 120kW of rack power without requiring a facilities modification. By Andy Patrizio May 15, 2024 3 mins Servers Data Center PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe