OpenAI, NVIDIA, AMD, Intel, Broadcom, and Microsoft have co-developed Multi-path Reliable Connections (MRC), a latest networking protocol designed to eliminate bottlenecks in AI supercomputers. By optimizing how data flows across Ethernet fabrics, MRC reduces tail latency and prevents training stalls, enabling the seamless scaling of next-generation LLMs.
For the uninitiated, training a frontier model isn’t just about having 100,000 H100s or B200s humming in a warehouse. It’s a massive orchestration problem. The real enemy isn’t raw compute power; it’s the “straggler.” In a synchronous training environment, every GPU must wait for the slowest packet of data to arrive before the next iteration of gradient descent can initiate. One dropped packet or one congested switch port doesn’t just slow down a single node—it brings the entire multi-billion-dollar cluster to a grinding halt.
This is the “bottleneck” the industry has been fighting. Until now, the solution was mostly a binary choice: stick with the lossy, unpredictable nature of standard Ethernet or pay the “NVIDIA tax” by using InfiniBand, a proprietary, low-latency fabric that works beautifully but locks you into a single vendor’s ecosystem.
MRC is the industry’s attempt to build a third way. It is AI-native Ethernet.
The Engineering of MRC: Killing the Tail Latency
To understand why MRC matters, you have to understand the failure of traditional TCP/IP in the context of LLM parameter scaling. Standard Ethernet is designed for the “best-effort” delivery of the internet—if a packet is lost, the system asks for it again. In a supercomputer, that re-transmission is a death sentence for efficiency. While IEEE standards have evolved, the overhead of the software stack remains too heavy.
MRC moves the reliability logic from the software layer down into the silicon. By leveraging Multi-pathing, the protocol doesn’t just send data along the shortest path; it spreads traffic across the entire fabric, dynamically routing around congestion in real-time. This is essentially a hardware-accelerated version of what we spot in sophisticated load balancers, but operating at the nanosecond scale required for GPU-to-GPU communication.
The protocol integrates tightly with RDMA (Remote Direct Memory Access), allowing one GPU to read the memory of another without involving the CPU. While RoCE v2 (RDMA over Converged Ethernet) attempted this, it was notoriously brittle, requiring painstaking manual tuning of Priority Flow Control (PFC) to avoid “congestion spreading”—a phenomenon where one clogged switch causes a ripple effect that freezes the whole network.
MRC replaces this fragility with a robust, hardware-offloaded reliability mechanism. It doesn’t just hope the packet arrives; it manages the delivery path with surgical precision.
The 30-Second Verdict: Hardware Comparison
| Feature | Standard Ethernet | InfiniBand | MRC-Enhanced Ethernet |
|---|---|---|---|
| Latency | High (Variable) | Ultra-Low (Consistent) | Low (Consistent) |
| Reliability | Software-based (TCP) | Hardware-based | Hardware-offloaded |
| Vendor Lock-in | None (Open) | High (NVIDIA) | Low (Consortium-based) |
| Scaling Cost | Low | Very High | Moderate |
Breaking the InfiniBand Monolith
Let’s talk about the macro-market dynamics. For years, NVIDIA’s InfiniBand has been the gold standard for AI clusters due to the fact that it is “lossless.” If you wanted to train a model with a trillion parameters, you bought InfiniBand. This gave NVIDIA a moat that extended far beyond the GPU itself; they owned the pipes, the switches, and the software that managed them.
By collaborating with AMD and Broadcom, OpenAI is effectively attempting to commoditize the AI fabric. If Ethernet can achieve “near-lossless” performance via MRC, the incentive to buy expensive, proprietary InfiniBand hardware vanishes. This is a strategic move to diversify the supply chain. Microsoft, in particular, cannot afford to be solely dependent on one vendor for the networking backbone of Azure AI.

This shift mirrors the transition we saw in the server market decades ago, moving from proprietary mainframes to x86 commodity hardware. We are now seeing the “commodity-ization” of the AI supercomputer interconnect.
“The move toward AI-native Ethernet isn’t just about cost; it’s about agility. When you can mix and match Broadcom switches with AMD GPUs and Intel NICs without sacrificing 20% of your compute efficiency to network jitter, the entire pace of LLM iteration accelerates.”
This quote reflects the sentiment currently echoing through the halls of high-performance computing (HPC) circles. The goal is a plug-and-play ecosystem where the fabric is invisible.
The Geopolitical Chessboard of Silicon
The inclusion of Intel and AMD in this protocol is the most telling detail. For the last three years, the “Chip War” has been framed as a battle of the GPUs. But the real war is being fought at the interconnect layer. If AMD’s Instinct accelerators can communicate across an MRC-enabled Ethernet fabric as efficiently as NVIDIA’s H100s do over InfiniBand, the performance gap narrows significantly.

this opens the door for a more robust open-source hardware movement. While the protocol is being developed by giants, the move toward an Ethernet-based standard makes it easier for third-party developers to build compatible networking gear. We are moving away from a “walled garden” and toward a “standardized city.”
However, don’t mistake this for NVIDIA’s defeat. NVIDIA is a founding member of the MRC effort. They are smart enough to realize that as the market scales to the “Gigascale” (clusters of millions of GPUs), even they cannot possibly manufacture enough proprietary cables and switches to satisfy global demand. By leading the Ethernet standard, they ensure that their Spectrum-X platforms remain the dominant implementation of that standard.
What This Means for Enterprise IT
- Reduced CapEx: Enterprises can leverage existing Ethernet expertise and hardware rather than hiring specialized InfiniBand engineers.
- Hybrid Clusters: The possibility of “heterogeneous” clusters—mixing GPU brands—becomes technically viable if they share a common, high-performance networking language.
- Faster Deployment: MRC reduces the “tuning” phase of cluster setup, which currently takes weeks of manual configuration to prevent packet loss.
The Final Analysis: Infrastructure as the New Alpha
We have spent the last two years obsessing over LLM parameter counts and token windows. But the real alpha in the AI race is now shifting toward infrastructure efficiency. The company that can train a model 10% faster because their network doesn’t stall is the company that reaches AGI first.
MRC is a tacit admission that the current networking stack was never designed for the massive, all-to-all communication patterns of transformer models. By rewriting the rules of the road, OpenAI and its partners are ensuring that the hardware can finally preserve up with the mathematics.
For those tracking the development of these systems, the next place to look is the open-source implementations of RDMA drivers. The real battle for the soul of the AI supercomputer will be fought in the kernel, where the handshake between the NPU and the network interface occurs. The “bottleneck” is being squeezed, and the result will be a surge in model scale that was previously unthinkable due to the physics of the wire.