Chelsio Looks to Close Ethernet-InfiniBand Gap

By Michael Feldman

January 24, 2013

This week Chelsio Communications unveiled its latest Ethernet adapter ASIC, which brings 40 gigabit speeds to its RDMA over TCP/IP (iWARP) portfolio. The fifth-generation silicon, dubbed Terminator T5, brings bandwidth and latency within spitting distance of FDR InfiniBand, and according to Chelsio, will actually outperform its IB competition on real-world HPC codes. According to Chelsio CEO and president Kianoosh Naghshineh, “the gap is essentially closed.”

Chelsio T5 ASIC sales are expected to ramp starting in Q2, while adapters based on the new silicon will roll out sometime later in the year. No pricing was given.

The T5, like its T4 predecessor, incorporates a TCP Offload Engine (TOE), iSCSI support, Fibre Channel over Ethernet (FCoE) and Network Address Translation (NAT) into hardware. But it’s the iWARP capability that is of special interest to HPC. While officially known as Internet Wide Area RDMA Protocol, iWARP is essentially RDMA over Ethernet, built on top of the ubiquitous TCP/IP protocol.

Chelsio T5 ASIC Architecture

And like all RDMA-based technology (which includes HPC’s go-to interconnect, InfiniBand), iWARP has the ability to bypass the CPU for data copies that tend to bottleneck the system. Done right, iWARP can offer performance on par with that of InfiniBand, and is eminently suitable for HPC clustering. And since it’s running atop TCP, iWARP is general-purpose enough to work in much larger and more heterogenous networks.

Better yet, since it’s supported by the OpenFabrics Enterprise Distribution (OFED), Linux applications written for InfiniBand can run seamlessly on iWARP-compatible gear. There’s no need to write codes specific to the protocol. The OpenFabrics group appears to be committed to maintaining this support in its software stack for the foreseeable future.

As an industry standard ratified by the Internet Engineering Task Force (IETF), iWARP is now backed by Intel, Broadcom, and Chelsio. Although Chelsio is the smallest of the three vendors, at this point it appears to be out ahead of its larger competitors. With the introduction of the T5, it is the only vendor that has married 40 Gig bandwidth and microsecond-level latencies to iWARP-style RDMA. Both Intel and Broadcom have 10 Gig implementations, but they are based on somewhat older technologies.

Intel, which inherited its 10GbE iWARP technology and expertise via its acquisition of NetEffect in 2008 hasn’t talked much about the product roadmap. However, along with Chelsio and Broadcom, Intel has been a driver in the most recent IETF extensions to the iWARP standard.

That suggests the chipmaker is going to move the NetEffect technology into the 40G realm (and beyond) at some point. And since Intel has outlined a network fabric strategy that integrates adapter logic into the CPU, it’s reasonable to assume that iWARP silicon could show up on x86 processors in the not-too-distant future.

None of that seems to worry Chelsio’s Naghshineh. According to him, their TCP offload technology leads the pack, which is probably why they sold 100,000 iWARP ports just in the last 12 months. What the market needs now though is a broader ecosystem, and that includes a reasonable number of iWARP providers committed to the technology. If Intel and Broadcom move ahead with their plans, that could provide the needed critical mass. “I’m very happy they are entering this market,” Naghshineh told HPCwire.

Thus far, most of Chelsio’s success has come from deployments in storage and virtualized servers, where the various network offloads supported by the adapter ASICs are used to save CPU cycles and boost performance. Chelsio’s penetration into the HPC space has been less sure – just a few university and commercial HPC installation to date. That’s due to a variety of factors, including delays in deployment of 10GbE technology overall and a perceived lack of iWARP performance relative to InfiniBand.

From Naghshineh’s perspective, the latter is just a marketing problem. There have been a number of studies that demonstrate even 10G iWARP performance is comparable to InfiniBand on typical HPC applications. One such analysis, performed by Chelsio, shows the its older T4 technology can perform as well or better than Mellanox’s FDR InfiniBand on typical MPI apps: LAMMPS (molecular dynamics), LS-DYNA (finite element analysis), WRF (weather forecasting), and HPL (Linpack).

Despite the FDR gear delivering four times the network bandwidth and half the latency of the Chelsio hardware, the study showed that the T4 iWARP implementation held its own across this application set. And the results seem to indicate that as the application scales up, the advantage starts to tilt in favor of iWARP. Since the newer T5 silicon brings the adapter bandwidth nearly up to FDR speeds (40Gbps versus 56Gbps) and latencies into the coveted sub-microsecond realm, Naghshineh expects the newer silicon to outperform the latest and greatest InfiniBand technology.

According to him, once they reach 100G iWARP in 2015, there will be no difference in performance between that and EDR InfiniBand, even at the hardware level. Naghshineh says that’s because the underlying SerDes (Serializer/Deserializer) architecture is converging across the different network technologies and that will become the common denominator determining performance.

Since Ethernet has the much larger ecosystem of switches, cables, optical modules, and software relative to InfiniBand, the economies of scale will naturally favor the high-volume solution, he maintains. And if performance and price are truly no longer differentiators between the two technologies, HPC users will come around. “InfiniBand has been a good solution to date, says Naghshineh. “It made sense to use it, but now the gap is essentially closed.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Quantum Companies D-Wave and Rigetti Again Face Stock Delisting

October 4, 2024

Both D-Wave (NYSE: QBTS) and Rigetti (Nasdaq: RGTI) are again facing stock delisting. This is a third time for D-Wave, which issued a press release today following notification by the SEC. Rigetti was notified of delisti Read more…

Alps Scientific Symposium Highlights AI’s Role in Tackling Science’s Biggest Challenges

October 4, 2024

ETH Zürich recently celebrated the launch of the AI-optimized “Alps” supercomputer with a scientific symposium focused on the future possibilities of scientific AI thanks to increased compute power and a flexible ar Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks. These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvidia GPUs). Recently, MLCommons introduced the results of its Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago today emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whatever physical processor they want, without making code changes, the Read more…

IBM Quantum Summit Evolves into Developer Conference

October 2, 2024

Instead of its usual quantum summit this year, IBM will hold its first IBM Quantum Developer Conference which the company is calling, “an exclusive, first-of-its-kind.” It’s planned as an in-person conference at th Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed that the company will release Falcon Shores as a GPU. The com Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks. These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvi Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago today emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whatever ph Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed t Read more…

How GenAI Will Impact Jobs In the Real World

September 30, 2024

There’s been a lot of fear, uncertainty, and doubt (FUD) about the potential for generative AI to take people’s jobs. The capability of large language model Read more…

IBM and NASA Launch Open-Source AI Model for Advanced Climate and Weather Research

September 25, 2024

IBM and NASA have developed a new AI foundation model for a wide range of climate and weather applications, with contributions from the Department of Energy’s Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Building the Quantum Economy — Chicago Style

September 24, 2024

Will there be regional winner in the global quantum economy sweepstakes? With visions of Silicon Valley’s iconic success in electronics and Boston/Cambridge� Read more…

How GPUs Are Embedded in the HPC Landscape

September 23, 2024

Grasping the basics of Graphics Processing Unit (GPU) architecture is crucial for understanding how these powerful processors function, particularly in high-per Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Granite Rapids HPC Benchmarks: I’m Thinking Intel Is Back (Updated)

September 25, 2024

Waiting is the hardest part. In the fall of 2023, HPCwire wrote about the new diverging Xeon processor strategy from Intel. Instead of a on-size-fits all approa Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Ansys Fluent® Adds AMD Instinct™ MI200 and MI300 Acceleration to Power CFD Simulations

September 23, 2024

Ansys Fluent® is well-known in the commercial computational fluid dynamics (CFD) space and is praised for its versatility as a general-purpose solver. Its impr Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Leading Solution Providers

Contributors

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

IBM Develops New Quantum Benchmarking Tool — Benchpress

September 26, 2024

Benchmarking is an important topic in quantum computing. There’s consensus it’s needed but opinions vary widely on how to go about it. Last week, IBM introd Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qu Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

US Implements Controls on Quantum Computing and other Technologies

September 27, 2024

Yesterday the Commerce Department announced export controls on quantum computing technologies as well as new controls for advanced semiconductors and additive Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire