Since 1986 - Covering the Fastest Computers in the World and the People Who Run Them

Language Flags
March 7, 2008

Ethernet, InfiniBand Musings

Michael Feldman

There seems to be a general consensus that the datacenter needs to settle on a unified network fabric. The question is, which one? Today, storage and cluster connectivity use some combination of Ethernet, InfiniBand and Fibre Channel, which creates a hodgepodge of cables, switches and host channel adapters. As readers of this publication are aware, both Ethernet and InfiniBand vendors have staked claim to unifying the datacenter on their blessed fabric. With the introduction of 10 GbE products and the upcoming Fibre Channel over Ethernet (FCoE) standards, Ethernet proponents claim cluster and storage unification is at hand. Meanwhile, the InfiniBand crowd is pushing their technology as the technically superior solution and the one that is available today.

Sun Microsystems seems to be getting impatient waiting for high performance Ethernet gear and FCoE standards, and has openly talked about going the InfiniBand route. In a recent EETimes article by Rick Merritt, Sun’s John Fowler (executive VP of the Systems Division) said the company would be introducing a set of products this year unified on InfiniBand. While Fowler admitted they’re continuing their involvement in the FCoE effort, he notes that Fibre Channel over InfiniBand should be available in 2008.

The strength of Ethernet — its ubiquity — is also it weakness. Changing or adding standards involves dragging the whole vendor community along. Look at how long the IEEE study group took just to decide to move forward on the 40GbE/100GbE standards. That process began in 2006 and standards aren’t expected to be completed until mid-2010. And while vendors may come up with proprietary versions of FCoE relatively soon, the standard is still being hashed out by the vendor and user communities. Not only does the process suffer from a “too many cooks in the kitchen” problem, but the cooks themselves are competing with one another.

Meanwhile, the more nimble InfiniBand community has been able to establish its high bandwidth, low-latency fabric as the standard for high performance interconnects. Mellanox has used its position as the only InfiniBand switch silicon vendor to set the pace. With DDR (20 Gbps) products now mainstream and QDR (40 Gbps) products due out at the end of this year, InfiniBand has left 10GbE behind from the standpoint of raw performance.

Increasingly, the InfiniBand vendors that have traditionally focused on the HPC market are looking at the broader enterprise marker. A couple of weeks ago at VMworld Europe, Mellanox announced that it is seeing demand for unified InfiniBand I/O in VMware environments. The company is reporting interest from sectors such as banking, managed hosting services, Web 2.0, insurance services and health care.

In the above mentioned EETimes piece, Sun Chief Architect Andy Bechtolsheim also points to the performance advantages of InfiniBand and, like Fowler, thinks InfiniBand may be the fabric that first swallows up Fibre Channel for storage connectivity. It’s a bit ironic that Bechtolsheim would be so enthusiastic about InfiniBand. A co-founder of Sun, he left the company in 1995 to start Granite Systems, a gigabit Ethernet switch vendor. A year later Cisco bought the company for $220 million. Now back at Sun, Bechtolsheim doesn’t seem to retain any nostalgia for his Ethernet roots. His most recent InfiniBand project, the 3,456 double data rate (DDR) Magnum switch, which powers the Sun Constellation “Ranger” supercomputer at TACC, is an example of Bechtolsheim’s enthusiasm for the technology.

The fact that he’s now bullish on InfiniBand should give unified fabric Ethernet enthusiasts something to think about. A shrewd investor, Bechtolsheim always seems to know which way the tech wind is blowing. As one of the original backers of Google, he turned a $100,000 investment into $1.5 billion. But like any accomplished investor, Bechtolsheim knows when to hedge his bets. He recently turned up as a backer for Arastra Inc., a vendor that has recently announced a high density, low-latency 10GbE switch for the datacenter.

In fact, almost everyone is expecting Ethernet to become the standard datacenter fabric … eventually. Even InfiniBand-loving Mellanox is hedging. The company’s recently announced ConnectX EN 10GigE NIC adapters are a nod to the ubiquity of Ethernet-based applications. And the ConnectX architecture is explicitly designed to support both InfiniBand and Ethernet connectivity on the same adapter. Essentially, the Mellanox approach is about unifying data traffic on a single I/O pipe rather than on a single protocol.

So is the fix in for Ethernet? Maybe not. As virtualization takes hold in the datacenter, the demand for ever more bandwidth and ever less latency is growing and some users can’t wait for 10GbE to get their act together. As John Lennon once said: “Life is what happens when you’re busy making plans” (or writing standards). Even mainstream technologies can collapse under their own weight (e.g., mainframes) or split up into a confusing array of variations (e.g., Unix). And let’s face it, if Ethernet gets pushed to the edge of the datacenter, civilization would manage to survive. But still, that scenario seems unlikely.

As I reported at the end of 2007, the Ethernet crowd thinks this year will be the one when 10GbE gets traction in the datacenter and even starts cutting away at InfiniBand’s dominance in supercomputing. But they’ve made that claim before. Sometimes, when I observe how the InfiniBand vendors are courting the OEMs, I’m reminded of the 2005 movie “Wedding Crashers.” Near the end of the story, lovestruck Owen Wilson pleads with Rachel McAdams: “I’m not standing here asking you to marry me. I’m just asking you not to marry him,” referring to McAdams’ obnoxious fiance. But while InfiniBand may be the HPC sentimental favorite, it might not be as fortunate as Wilson in getting that Hollywood ending.

As always, comments about HPCwire are welcomed and encouraged. Write to me, Michael Feldman, at editor@hpcwire.com.