UALink Debuts Its First AI Interconnect Spec – Usable In Just 18 Short Months

The Ultra Accelerator Link Consortium has delivered its first GPU interconnect specification: UALink 200G 1.0.

The Consortium (UAC) was formed in May 2024 by a group of vendors – among them AMD, AWS, Broadcom, Cisco, Google, HPE, Intel, Meta, Microsoft, and Astera Labs – who think the world needs an open alternative to Nvidia's NVLink tech that allows creation of the networked GPU clusters needed to run AI workloads at scale.

Members aren't just advancing the cause of open standards. Nvidia's networking business won over $13 billion in revenue in its last financial year, and the GPU giant has signaled its intention to grow that business. UALink's members fancy creating a cheaper alternative they can control and deploy themselves at hyperscale, or profit from by creating hardware the rest of us buy.

They also think the world is ready for a networking standard that can be applied to GPUs from multiple vendors rather than requiring users to create network silos dedicated to each accelerator vendor.

To satisfy those goals, the UAC also wants to work over the Ethernet networks most orgs already operate.

UALink 200G 1.0, as the name implies, enables a 200 Gbps (gigabits per second) connection to an accelerator. It can also quadruple that speed by allowing four connections to each GPU.

The spec allows creation of compute pods packing 1,024 accelerators and achieving what the consortium describes as "the same raw speed as Ethernet with the latency of PCIe switches."

All while consuming somewhere between a third and a half of a typical Ethernet network.

That's a lot to get done in under a year, but UAC didn't start from scratch. Chair and director Kurtis Bowman told The Register the spec draws very heavily on AMD's existing Infinity Fabric product.

"We were able to build on that [Infinity Fabric]," he told The Register, but also used tech from other UAC members who have cooked their own networking stacks to address their own needs.

"Intel, Google, and Microsoft said 'we have challenges in our datacenters and we need you to address that,'" Bowman said.

He admitted it will be around 18 months before compliant hardware goes on sale but thinks that's six months less than is typically required to turn a spec into product. Bowman thinks the likes of HPE, Dell, and Lenovo will adopt the spec and deliver AI solutions that employ it, as will the likes of Broadcom and Synopsys as they create custom accelerators for hyperscale customers.

Work on a second spec is already underway to take advantage of 400G Ethernet variants as they go mainstream. ®

RECENT NEWS

From Chip War To Cloud War: The Next Frontier In Global Tech Competition

The global chip war, characterized by intense competition among nations and corporations for supremacy in semiconductor ... Read more

The High Stakes Of Tech Regulation: Security Risks And Market Dynamics

The influence of tech giants in the global economy continues to grow, raising crucial questions about how to balance sec... Read more

The Tyranny Of Instagram Interiors: Why It's Time To Break Free From Algorithm-Driven Aesthetics

Instagram has become a dominant force in shaping interior design trends, offering a seemingly endless stream of inspirat... Read more

The Data Crunch In AI: Strategies For Sustainability

Exploring solutions to the imminent exhaustion of internet data for AI training.As the artificial intelligence (AI) indu... Read more

Google Abandons Four-Year Effort To Remove Cookies From Chrome Browser

After four years of dedicated effort, Google has decided to abandon its plan to remove third-party cookies from its Chro... Read more

LinkedIn Embraces AI And Gamification To Drive User Engagement And Revenue

In an effort to tackle slowing revenue growth and enhance user engagement, LinkedIn is turning to artificial intelligenc... Read more