Convergence of Ultra Ethernet, CXL, and PCIe Toward a Rack-Scale Computer Fabric
Over the next several years, the boundaries between device interconnects and datacenter networks are likely to blur. Technologies such as PCI Express, Compute Express Link, and the emerging Ultra Ethernet architecture are evolving toward similar goals: moving large amounts of data between compute elements with minimal software overhead and with latency characteristics closer to memory access than to traditional networking. As AI infrastructure grows in scale, these technologies are increasingly being positioned not as separate layers but as components of a unified communication fabric that spans an entire rack.
The shift from servers to rack-scale systems

Large AI workloads, however, behave differently. Training large models involves thousands of GPUs exchanging gradients and tensors continuously. The boundaries between servers begin to matter less than the total pool of compute, memory, and accelerator resources available across the cluster. As a result, system designers increasingly treat a rack—or even a row of racks—as a single logical compute unit. The interconnect technologies that link components inside a server therefore begin to resemble those used between servers.
Complementary roles of the three fabrics

In a future rack-scale architecture, these technologies may form a layered continuum rather than distinct systems. PCIe would continue to serve as the immediate device interconnect within a node. CXL would enable memory sharing and pooling across multiple processors and accelerators within the rack. Ultra Ethernet would extend this communication across nodes while preserving many of the same semantics—direct memory placement, low latency messaging, and hardware-managed communication.
Convergence around memory semantics

Ultra Ethernet builds on this trend by optimizing Ethernet for large-scale accelerator communication. Instead of treating the network purely as a packet transport, it begins to resemble a distributed memory interconnect in which devices communicate through memory-like operations. This makes the conceptual boundary between CXL memory transactions and network transfers increasingly narrow.
Latency and determinism requirements

PCIe already provides deterministic communication within a server. CXL aims to extend deterministic access to shared memory pools. Ultra Ethernet introduces mechanisms such as credit-based flow control, link-level retries, and hardware message matching to achieve more predictable behavior in large clusters. As these features evolve, the communication characteristics of datacenter networks begin to resemble those of device interconnect fabrics.
Hardware offload and accelerator-centric networking

As a result, the boundaries between PCIe controllers, CXL interfaces, and high-performance NICs are beginning to blur. Future silicon designs may integrate features traditionally associated with all three technologies into unified I/O subsystems capable of handling device communication, memory sharing, and cluster networking through a common architecture.
Resource disaggregation and pooling

Together, these technologies enable a model in which resources within a rack are no longer fixed to a single server but can be allocated dynamically across workloads. A rack may contain pools of CPUs, GPUs, and memory modules that behave collectively as a single composable computer.
Toward the rack as the system

The convergence of PCIe, CXL, and Ultra Ethernet reflects this architectural shift. Each technology contributes elements required to make a rack-scale computer possible: high-speed device connectivity, shared memory semantics, and scalable inter-node communication. As these capabilities continue to align, the distinction between internal system interconnects and datacenter networks may gradually disappear, replaced by a unified fabric spanning the entire rack.

